[ 488.924090] env[61868]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=61868) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 488.924468] env[61868]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=61868) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 488.924532] env[61868]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=61868) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 488.924861] env[61868]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 489.018714] env[61868]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=61868) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 489.028304] env[61868]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.009s {{(pid=61868) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 489.634758] env[61868]: INFO nova.virt.driver [None req-28156aaf-e4e8-4ace-931a-93dc0df359db None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 489.712162] env[61868]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 489.712324] env[61868]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 489.712421] env[61868]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=61868) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 492.728121] env[61868]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-d5d00a8e-1d55-4bf0-94f6-84900d0b3935 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.744110] env[61868]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=61868) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 492.744325] env[61868]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-7767b71d-b7ae-4ff9-8c0f-b4e7d8ecd3ae {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.771771] env[61868]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 7cc88. [ 492.771992] env[61868]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.060s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 492.772461] env[61868]: INFO nova.virt.vmwareapi.driver [None req-28156aaf-e4e8-4ace-931a-93dc0df359db None None] VMware vCenter version: 7.0.3 [ 492.775803] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a9e25eb-060e-4656-8459-bfd9e5b74ad3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.792996] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-849120bf-7a4c-4d05-a6d6-d3a81fd9d9e6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.798815] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ff8449b-c8f5-4a51-9eae-97886069d01e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.805371] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51bbcf99-22e1-47af-ae2b-c26a6d9f6271 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.818036] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-261f9806-9fd3-4822-a9d9-349983f76927 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.823884] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8168ced1-1078-4121-b683-b8ea9194f438 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.853707] env[61868]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-ce176077-d969-470d-a2f2-11dd60e84dd1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.858678] env[61868]: DEBUG nova.virt.vmwareapi.driver [None req-28156aaf-e4e8-4ace-931a-93dc0df359db None None] Extension org.openstack.compute already exists. {{(pid=61868) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:226}} [ 492.861441] env[61868]: INFO nova.compute.provider_config [None req-28156aaf-e4e8-4ace-931a-93dc0df359db None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 493.364679] env[61868]: DEBUG nova.context [None req-28156aaf-e4e8-4ace-931a-93dc0df359db None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),0f616958-a9ac-4e3e-a87e-3b9f7af08314(cell1) {{(pid=61868) load_cells /opt/stack/nova/nova/context.py:464}} [ 493.367225] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 493.367530] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 493.368516] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 493.369108] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] Acquiring lock "0f616958-a9ac-4e3e-a87e-3b9f7af08314" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 493.369363] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] Lock "0f616958-a9ac-4e3e-a87e-3b9f7af08314" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 493.370776] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] Lock "0f616958-a9ac-4e3e-a87e-3b9f7af08314" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 493.391409] env[61868]: INFO dbcounter [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] Registered counter for database nova_cell0 [ 493.399338] env[61868]: INFO dbcounter [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] Registered counter for database nova_cell1 [ 493.402692] env[61868]: DEBUG oslo_db.sqlalchemy.engines [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61868) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 493.403338] env[61868]: DEBUG oslo_db.sqlalchemy.engines [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61868) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 493.408020] env[61868]: ERROR nova.db.main.api [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 493.408020] env[61868]: result = function(*args, **kwargs) [ 493.408020] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 493.408020] env[61868]: return func(*args, **kwargs) [ 493.408020] env[61868]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 493.408020] env[61868]: result = fn(*args, **kwargs) [ 493.408020] env[61868]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 493.408020] env[61868]: return f(*args, **kwargs) [ 493.408020] env[61868]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 493.408020] env[61868]: return db.service_get_minimum_version(context, binaries) [ 493.408020] env[61868]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 493.408020] env[61868]: _check_db_access() [ 493.408020] env[61868]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 493.408020] env[61868]: stacktrace = ''.join(traceback.format_stack()) [ 493.408020] env[61868]: [ 493.408847] env[61868]: ERROR nova.db.main.api [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 493.408847] env[61868]: result = function(*args, **kwargs) [ 493.408847] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 493.408847] env[61868]: return func(*args, **kwargs) [ 493.408847] env[61868]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 493.408847] env[61868]: result = fn(*args, **kwargs) [ 493.408847] env[61868]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 493.408847] env[61868]: return f(*args, **kwargs) [ 493.408847] env[61868]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 493.408847] env[61868]: return db.service_get_minimum_version(context, binaries) [ 493.408847] env[61868]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 493.408847] env[61868]: _check_db_access() [ 493.408847] env[61868]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 493.408847] env[61868]: stacktrace = ''.join(traceback.format_stack()) [ 493.408847] env[61868]: [ 493.409486] env[61868]: WARNING nova.objects.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 493.409486] env[61868]: WARNING nova.objects.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] Failed to get minimum service version for cell 0f616958-a9ac-4e3e-a87e-3b9f7af08314 [ 493.409817] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] Acquiring lock "singleton_lock" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 493.410324] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] Acquired lock "singleton_lock" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 493.410562] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] Releasing lock "singleton_lock" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 493.410820] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] Full set of CONF: {{(pid=61868) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 493.410963] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] ******************************************************************************** {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 493.411117] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] Configuration options gathered from: {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 493.411257] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 493.411455] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 493.411582] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] ================================================================================ {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 493.411786] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] allow_resize_to_same_host = True {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.411950] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] arq_binding_timeout = 300 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.412095] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] backdoor_port = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.412225] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] backdoor_socket = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.412388] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] block_device_allocate_retries = 60 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.412547] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] block_device_allocate_retries_interval = 3 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.412714] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cert = self.pem {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.412878] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.413058] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] compute_monitors = [] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.413234] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] config_dir = [] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.413443] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] config_drive_format = iso9660 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.413533] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.413697] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] config_source = [] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.413863] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] console_host = devstack {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.414038] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] control_exchange = nova {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.414203] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cpu_allocation_ratio = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.414365] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] daemon = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.414528] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] debug = True {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.414686] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] default_access_ip_network_name = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.414850] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] default_availability_zone = nova {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.415016] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] default_ephemeral_format = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.415187] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] default_green_pool_size = 1000 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.415426] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.415591] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] default_schedule_zone = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.415750] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] disk_allocation_ratio = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.415929] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] enable_new_services = True {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.416142] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] enabled_apis = ['osapi_compute'] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.416314] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] enabled_ssl_apis = [] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.416476] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] flat_injected = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.416633] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] force_config_drive = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.416789] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] force_raw_images = True {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.416979] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] graceful_shutdown_timeout = 5 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.417160] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] heal_instance_info_cache_interval = 60 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.417381] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] host = cpu-1 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.417554] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.417714] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] initial_disk_allocation_ratio = 1.0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.417874] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] initial_ram_allocation_ratio = 1.0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.418139] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.418315] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] instance_build_timeout = 0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.418476] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] instance_delete_interval = 300 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.418644] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] instance_format = [instance: %(uuid)s] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.418810] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] instance_name_template = instance-%08x {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.418989] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] instance_usage_audit = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.419206] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] instance_usage_audit_period = month {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.419381] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.419549] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] instances_path = /opt/stack/data/nova/instances {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.419716] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] internal_service_availability_zone = internal {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.419875] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] key = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.420047] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] live_migration_retry_count = 30 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.420224] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] log_color = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.420389] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] log_config_append = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.420555] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.420716] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] log_dir = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.420872] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] log_file = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.420999] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] log_options = True {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.421179] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] log_rotate_interval = 1 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.421350] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] log_rotate_interval_type = days {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.421518] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] log_rotation_type = none {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.421707] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.421779] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.421937] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.422140] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.422277] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.422445] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] long_rpc_timeout = 1800 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.422604] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] max_concurrent_builds = 10 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.422763] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] max_concurrent_live_migrations = 1 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.422921] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] max_concurrent_snapshots = 5 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.423089] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] max_local_block_devices = 3 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.423251] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] max_logfile_count = 30 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.423408] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] max_logfile_size_mb = 200 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.423566] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] maximum_instance_delete_attempts = 5 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.423732] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] metadata_listen = 0.0.0.0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.423902] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] metadata_listen_port = 8775 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.424077] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] metadata_workers = 2 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.424241] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] migrate_max_retries = -1 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.424407] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] mkisofs_cmd = genisoimage {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.424614] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] my_block_storage_ip = 10.180.1.21 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.424745] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] my_ip = 10.180.1.21 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.424908] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] network_allocate_retries = 0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.425123] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.425308] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] osapi_compute_listen = 0.0.0.0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.425473] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] osapi_compute_listen_port = 8774 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.425640] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] osapi_compute_unique_server_name_scope = {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.425808] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] osapi_compute_workers = 2 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.425999] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] password_length = 12 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.426187] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] periodic_enable = True {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.426350] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] periodic_fuzzy_delay = 60 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.426519] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] pointer_model = usbtablet {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.426686] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] preallocate_images = none {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.426867] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] publish_errors = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.426976] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] pybasedir = /opt/stack/nova {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.427147] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] ram_allocation_ratio = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.427308] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] rate_limit_burst = 0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.427476] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] rate_limit_except_level = CRITICAL {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.427637] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] rate_limit_interval = 0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.427794] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] reboot_timeout = 0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.427951] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] reclaim_instance_interval = 0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.428148] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] record = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.428322] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] reimage_timeout_per_gb = 60 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.428488] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] report_interval = 120 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.428647] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] rescue_timeout = 0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.428806] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] reserved_host_cpus = 0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.428965] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] reserved_host_disk_mb = 0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.429136] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] reserved_host_memory_mb = 512 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.429295] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] reserved_huge_pages = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.429452] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] resize_confirm_window = 0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.429611] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] resize_fs_using_block_device = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.429768] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] resume_guests_state_on_host_boot = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.429933] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.430105] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] rpc_response_timeout = 60 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.430269] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] run_external_periodic_tasks = True {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.430437] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] running_deleted_instance_action = reap {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.430595] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] running_deleted_instance_poll_interval = 1800 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.430750] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] running_deleted_instance_timeout = 0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.430905] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] scheduler_instance_sync_interval = 120 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.431078] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] service_down_time = 720 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.431250] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] servicegroup_driver = db {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.431403] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] shell_completion = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.431558] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] shelved_offload_time = 0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.431714] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] shelved_poll_interval = 3600 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.431877] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] shutdown_timeout = 0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.432046] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] source_is_ipv6 = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.432210] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] ssl_only = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.432457] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.432622] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] sync_power_state_interval = 600 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.432781] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] sync_power_state_pool_size = 1000 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.432946] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] syslog_log_facility = LOG_USER {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.433117] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] tempdir = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.433278] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] timeout_nbd = 10 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.433442] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] transport_url = **** {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.433660] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] update_resources_interval = 0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.433756] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] use_cow_images = True {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.433910] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] use_eventlog = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.434078] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] use_journal = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.434241] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] use_json = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.434397] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] use_rootwrap_daemon = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.434553] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] use_stderr = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.434708] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] use_syslog = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.434861] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vcpu_pin_set = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.435038] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vif_plugging_is_fatal = True {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.435212] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vif_plugging_timeout = 300 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.435377] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] virt_mkfs = [] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.435537] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] volume_usage_poll_interval = 0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.435696] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] watch_log_file = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.435866] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] web = /usr/share/spice-html5 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.436136] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.436321] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.436499] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.436672] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_concurrency.disable_process_locking = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.437305] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.437507] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.437684] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.437863] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.438076] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.438270] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.438456] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api.auth_strategy = keystone {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.438628] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api.compute_link_prefix = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.438804] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.438994] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api.dhcp_domain = novalocal {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.439209] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api.enable_instance_password = True {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.439385] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api.glance_link_prefix = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.439556] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.439730] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.439892] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api.instance_list_per_project_cells = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.440066] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api.list_records_by_skipping_down_cells = True {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.440237] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api.local_metadata_per_cell = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.440407] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api.max_limit = 1000 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.440592] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api.metadata_cache_expiration = 15 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.440750] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api.neutron_default_tenant_id = default {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.440921] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api.response_validation = warn {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.441104] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api.use_neutron_default_nets = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.441279] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.441443] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.441610] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.441782] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.441953] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api.vendordata_dynamic_targets = [] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.442129] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api.vendordata_jsonfile_path = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.442315] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.442511] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.backend = dogpile.cache.memcached {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.442679] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.backend_argument = **** {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.442851] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.config_prefix = cache.oslo {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.443033] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.dead_timeout = 60.0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.443205] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.debug_cache_backend = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.443366] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.enable_retry_client = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.443526] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.enable_socket_keepalive = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.443703] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.enabled = True {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.443864] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.enforce_fips_mode = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.444037] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.expiration_time = 600 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.444208] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.hashclient_retry_attempts = 2 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.444374] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.hashclient_retry_delay = 1.0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.444536] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.memcache_dead_retry = 300 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.444694] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.memcache_password = **** {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.444859] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.445033] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.445204] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.memcache_pool_maxsize = 10 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.445367] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.445530] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.memcache_sasl_enabled = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.445709] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.445876] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.memcache_socket_timeout = 1.0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.446049] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.memcache_username = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.446221] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.proxies = [] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.446386] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.redis_db = 0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.446546] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.redis_password = **** {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.446716] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.redis_sentinel_service_name = mymaster {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.446908] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.447098] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.redis_server = localhost:6379 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.447279] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.redis_socket_timeout = 1.0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.447443] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.redis_username = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.447606] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.retry_attempts = 2 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.447771] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.retry_delay = 0.0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.447932] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.socket_keepalive_count = 1 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.448140] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.socket_keepalive_idle = 1 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.448315] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.socket_keepalive_interval = 1 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.448477] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.tls_allowed_ciphers = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.448636] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.tls_cafile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.448794] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.tls_certfile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.448987] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.tls_enabled = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.449188] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cache.tls_keyfile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.449366] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cinder.auth_section = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.449541] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cinder.auth_type = password {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.449703] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cinder.cafile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.449877] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cinder.catalog_info = volumev3::publicURL {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.450047] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cinder.certfile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.450217] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cinder.collect_timing = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.450382] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cinder.cross_az_attach = True {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.450544] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cinder.debug = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.450701] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cinder.endpoint_template = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.450864] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cinder.http_retries = 3 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.451039] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cinder.insecure = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.451206] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cinder.keyfile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.451376] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cinder.os_region_name = RegionOne {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.451539] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cinder.split_loggers = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.451697] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cinder.timeout = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.451869] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.452111] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] compute.cpu_dedicated_set = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.452298] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] compute.cpu_shared_set = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.452471] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] compute.image_type_exclude_list = [] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.452639] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.452806] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] compute.max_concurrent_disk_ops = 0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.452972] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] compute.max_disk_devices_to_attach = -1 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.453152] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.453325] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.453487] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] compute.resource_provider_association_refresh = 300 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.453650] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.453811] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] compute.shutdown_retry_interval = 10 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.453994] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.454192] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] conductor.workers = 2 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.454375] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] console.allowed_origins = [] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.454538] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] console.ssl_ciphers = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.454709] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] console.ssl_minimum_version = default {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.454878] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] consoleauth.enforce_session_timeout = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.455085] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] consoleauth.token_ttl = 600 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.455267] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cyborg.cafile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.455430] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cyborg.certfile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.455598] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cyborg.collect_timing = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.455759] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cyborg.connect_retries = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.455921] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cyborg.connect_retry_delay = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.456094] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cyborg.endpoint_override = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.456265] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cyborg.insecure = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.456424] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cyborg.keyfile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.456586] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cyborg.max_version = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.456744] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cyborg.min_version = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.456903] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cyborg.region_name = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.457074] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cyborg.retriable_status_codes = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.457237] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cyborg.service_name = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.457405] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cyborg.service_type = accelerator {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.457566] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cyborg.split_loggers = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.457723] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cyborg.status_code_retries = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.457879] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cyborg.status_code_retry_delay = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.458086] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cyborg.timeout = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.458291] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.458458] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] cyborg.version = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.458643] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] database.backend = sqlalchemy {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.458823] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] database.connection = **** {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.458983] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] database.connection_debug = 0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.459172] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] database.connection_parameters = {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.459339] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] database.connection_recycle_time = 3600 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.459503] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] database.connection_trace = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.459667] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] database.db_inc_retry_interval = True {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.459830] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] database.db_max_retries = 20 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.459992] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] database.db_max_retry_interval = 10 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.460168] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] database.db_retry_interval = 1 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.460414] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] database.max_overflow = 50 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.460498] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] database.max_pool_size = 5 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.460653] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] database.max_retries = 10 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.460822] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.460982] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] database.mysql_wsrep_sync_wait = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.461188] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] database.pool_timeout = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.461362] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] database.retry_interval = 10 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.461522] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] database.slave_connection = **** {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.461687] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] database.sqlite_synchronous = True {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.461850] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] database.use_db_reconnect = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.462041] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api_database.backend = sqlalchemy {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.462221] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api_database.connection = **** {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.462388] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api_database.connection_debug = 0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.462555] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api_database.connection_parameters = {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.462717] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api_database.connection_recycle_time = 3600 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.462878] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api_database.connection_trace = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.463050] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api_database.db_inc_retry_interval = True {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.463218] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api_database.db_max_retries = 20 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.463379] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api_database.db_max_retry_interval = 10 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.463538] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api_database.db_retry_interval = 1 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.463696] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api_database.max_overflow = 50 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.463855] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api_database.max_pool_size = 5 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.464022] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api_database.max_retries = 10 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.464227] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.464394] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.464555] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api_database.pool_timeout = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.464717] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api_database.retry_interval = 10 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.464873] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api_database.slave_connection = **** {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.465045] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] api_database.sqlite_synchronous = True {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.465227] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] devices.enabled_mdev_types = [] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.465404] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.465576] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] ephemeral_storage_encryption.default_format = luks {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.465739] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] ephemeral_storage_encryption.enabled = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.465902] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.466086] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] glance.api_servers = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.466256] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] glance.cafile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.466419] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] glance.certfile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.466585] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] glance.collect_timing = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.466745] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] glance.connect_retries = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.466903] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] glance.connect_retry_delay = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.467092] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] glance.debug = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.467283] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] glance.default_trusted_certificate_ids = [] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.467452] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] glance.enable_certificate_validation = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.467617] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] glance.enable_rbd_download = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.467778] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] glance.endpoint_override = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.467945] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] glance.insecure = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.468158] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] glance.keyfile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.468326] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] glance.max_version = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.468486] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] glance.min_version = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.468650] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] glance.num_retries = 3 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.468821] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] glance.rbd_ceph_conf = {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.468984] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] glance.rbd_connect_timeout = 5 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.469170] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] glance.rbd_pool = {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.469338] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] glance.rbd_user = {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.469497] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] glance.region_name = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.469658] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] glance.retriable_status_codes = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.469818] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] glance.service_name = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.469985] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] glance.service_type = image {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.470186] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] glance.split_loggers = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.470370] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] glance.status_code_retries = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.470532] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] glance.status_code_retry_delay = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.470692] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] glance.timeout = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.470876] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.471055] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] glance.verify_glance_signatures = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.471222] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] glance.version = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.471392] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] guestfs.debug = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.471560] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] mks.enabled = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.471914] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.472118] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] image_cache.manager_interval = 2400 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.472294] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] image_cache.precache_concurrency = 1 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.472464] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] image_cache.remove_unused_base_images = True {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.472633] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.472802] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.472980] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] image_cache.subdirectory_name = _base {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.473191] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] ironic.api_max_retries = 60 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.473374] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] ironic.api_retry_interval = 2 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.473537] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] ironic.auth_section = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.473702] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] ironic.auth_type = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.473864] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] ironic.cafile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.474072] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] ironic.certfile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.474200] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] ironic.collect_timing = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.474363] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] ironic.conductor_group = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.474522] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] ironic.connect_retries = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.474680] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] ironic.connect_retry_delay = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.474840] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] ironic.endpoint_override = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.475009] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] ironic.insecure = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.475178] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] ironic.keyfile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.475334] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] ironic.max_version = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.475490] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] ironic.min_version = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.475654] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] ironic.peer_list = [] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.475811] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] ironic.region_name = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.476014] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] ironic.retriable_status_codes = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.476200] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] ironic.serial_console_state_timeout = 10 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.476364] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] ironic.service_name = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.476536] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] ironic.service_type = baremetal {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.476696] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] ironic.shard = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.476861] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] ironic.split_loggers = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.477028] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] ironic.status_code_retries = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.477194] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] ironic.status_code_retry_delay = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.477724] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] ironic.timeout = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.477724] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.477724] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] ironic.version = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.477846] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.478050] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] key_manager.fixed_key = **** {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.478254] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.478419] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] barbican.barbican_api_version = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.478580] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] barbican.barbican_endpoint = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.478752] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] barbican.barbican_endpoint_type = public {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.478922] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] barbican.barbican_region_name = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.479108] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] barbican.cafile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.479274] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] barbican.certfile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.479440] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] barbican.collect_timing = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.479605] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] barbican.insecure = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.479765] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] barbican.keyfile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.479931] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] barbican.number_of_retries = 60 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.480109] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] barbican.retry_delay = 1 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.480274] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] barbican.send_service_user_token = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.480436] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] barbican.split_loggers = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.480591] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] barbican.timeout = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.480751] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] barbican.verify_ssl = True {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.480909] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] barbican.verify_ssl_path = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.481088] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] barbican_service_user.auth_section = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.481255] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] barbican_service_user.auth_type = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.481415] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] barbican_service_user.cafile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.481573] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] barbican_service_user.certfile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.481735] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] barbican_service_user.collect_timing = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.481896] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] barbican_service_user.insecure = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.482095] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] barbican_service_user.keyfile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.482268] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] barbican_service_user.split_loggers = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.482428] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] barbican_service_user.timeout = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.482594] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vault.approle_role_id = **** {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.482754] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vault.approle_secret_id = **** {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.482923] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vault.kv_mountpoint = secret {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.483095] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vault.kv_path = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.483266] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vault.kv_version = 2 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.483427] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vault.namespace = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.483588] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vault.root_token_id = **** {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.483747] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vault.ssl_ca_crt_file = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.483914] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vault.timeout = 60.0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.484089] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vault.use_ssl = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.484265] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.484435] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] keystone.auth_section = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.484636] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] keystone.auth_type = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.484803] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] keystone.cafile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.484978] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] keystone.certfile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.485189] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] keystone.collect_timing = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.485363] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] keystone.connect_retries = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.485565] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] keystone.connect_retry_delay = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.485742] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] keystone.endpoint_override = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.485916] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] keystone.insecure = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.486084] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] keystone.keyfile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.486249] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] keystone.max_version = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.486406] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] keystone.min_version = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.486565] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] keystone.region_name = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.486725] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] keystone.retriable_status_codes = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.486884] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] keystone.service_name = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.487067] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] keystone.service_type = identity {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.487235] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] keystone.split_loggers = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.487406] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] keystone.status_code_retries = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.487555] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] keystone.status_code_retry_delay = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.487711] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] keystone.timeout = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.487892] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.488106] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] keystone.version = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.488317] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.connection_uri = {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.488482] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.cpu_mode = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.488652] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.cpu_model_extra_flags = [] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.488822] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.cpu_models = [] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.489009] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.cpu_power_governor_high = performance {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.489183] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.cpu_power_governor_low = powersave {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.489348] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.cpu_power_management = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.489520] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.489684] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.device_detach_attempts = 8 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.489847] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.device_detach_timeout = 20 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.490029] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.disk_cachemodes = [] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.490182] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.disk_prefix = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.490348] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.enabled_perf_events = [] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.490511] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.file_backed_memory = 0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.490677] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.gid_maps = [] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.490836] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.hw_disk_discard = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.490994] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.hw_machine_type = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.491217] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.images_rbd_ceph_conf = {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.491393] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.491561] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.491732] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.images_rbd_glance_store_name = {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.491903] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.images_rbd_pool = rbd {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.492088] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.images_type = default {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.492254] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.images_volume_group = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.492418] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.inject_key = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.492580] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.inject_partition = -2 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.492740] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.inject_password = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.492903] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.iscsi_iface = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.493076] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.iser_use_multipath = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.493244] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.live_migration_bandwidth = 0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.493450] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.493564] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.live_migration_downtime = 500 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.493724] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.493885] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.494071] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.live_migration_inbound_addr = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.494286] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.494424] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.live_migration_permit_post_copy = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.494589] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.live_migration_scheme = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.494765] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.live_migration_timeout_action = abort {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.494932] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.live_migration_tunnelled = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.495110] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.live_migration_uri = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.495281] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.live_migration_with_native_tls = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.495441] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.max_queues = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.495604] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.495839] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.496039] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.nfs_mount_options = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.496358] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.496535] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.496701] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.num_iser_scan_tries = 5 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.496864] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.num_memory_encrypted_guests = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.497043] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.497244] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.num_pcie_ports = 0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.497419] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.num_volume_scan_tries = 5 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.497627] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.pmem_namespaces = [] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.497744] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.quobyte_client_cfg = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.498075] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.498261] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.rbd_connect_timeout = 5 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.498432] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.498598] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.498763] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.rbd_secret_uuid = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.498924] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.rbd_user = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.499104] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.499283] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.remote_filesystem_transport = ssh {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.499444] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.rescue_image_id = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.499604] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.rescue_kernel_id = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.499764] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.rescue_ramdisk_id = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.499934] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.500118] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.rx_queue_size = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.500312] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.smbfs_mount_options = {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.500594] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.500770] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.snapshot_compression = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.500934] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.snapshot_image_format = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.501194] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.501379] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.sparse_logical_volumes = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.501547] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.swtpm_enabled = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.501720] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.swtpm_group = tss {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.501892] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.swtpm_user = tss {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.502084] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.sysinfo_serial = unique {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.502253] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.tb_cache_size = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.502413] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.tx_queue_size = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.502579] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.uid_maps = [] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.502742] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.use_virtio_for_bridges = True {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.502910] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.virt_type = kvm {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.503094] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.volume_clear = zero {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.503289] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.volume_clear_size = 0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.503464] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.volume_use_multipath = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.503624] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.vzstorage_cache_path = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.503796] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.503965] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.vzstorage_mount_group = qemu {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.504149] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.vzstorage_mount_opts = [] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.504318] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.504593] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.504768] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.vzstorage_mount_user = stack {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.504935] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.505124] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] neutron.auth_section = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.505302] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] neutron.auth_type = password {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.505465] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] neutron.cafile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.505626] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] neutron.certfile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.505789] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] neutron.collect_timing = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.505980] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] neutron.connect_retries = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.506172] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] neutron.connect_retry_delay = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.506347] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] neutron.default_floating_pool = public {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.506506] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] neutron.endpoint_override = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.506668] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] neutron.extension_sync_interval = 600 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.506829] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] neutron.http_retries = 3 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.506990] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] neutron.insecure = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.507170] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] neutron.keyfile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.507332] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] neutron.max_version = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.507504] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.507662] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] neutron.min_version = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.507830] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] neutron.ovs_bridge = br-int {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.508034] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] neutron.physnets = [] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.508218] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] neutron.region_name = RegionOne {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.508381] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] neutron.retriable_status_codes = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.508551] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] neutron.service_metadata_proxy = True {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.508717] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] neutron.service_name = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.508887] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] neutron.service_type = network {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.509137] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] neutron.split_loggers = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.509338] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] neutron.status_code_retries = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.509545] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] neutron.status_code_retry_delay = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.509735] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] neutron.timeout = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.509931] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.510111] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] neutron.version = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.510315] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] notifications.bdms_in_notifications = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.510470] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] notifications.default_level = INFO {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.510646] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] notifications.notification_format = unversioned {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.510808] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] notifications.notify_on_state_change = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.510985] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.511181] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] pci.alias = [] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.511373] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] pci.device_spec = [] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.511553] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] pci.report_in_placement = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.511751] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.auth_section = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.511993] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.auth_type = password {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.512214] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.512384] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.cafile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.512546] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.certfile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.512726] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.collect_timing = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.512904] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.connect_retries = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.513137] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.connect_retry_delay = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.513317] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.default_domain_id = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.513488] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.default_domain_name = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.513649] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.domain_id = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.513825] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.domain_name = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.513965] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.endpoint_override = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.514145] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.insecure = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.514304] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.keyfile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.514500] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.max_version = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.514613] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.min_version = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.514780] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.password = **** {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.514942] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.project_domain_id = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.515152] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.project_domain_name = Default {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.515329] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.project_id = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.515503] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.project_name = service {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.515676] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.region_name = RegionOne {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.515838] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.retriable_status_codes = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.516037] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.service_name = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.516220] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.service_type = placement {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.516384] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.split_loggers = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.516542] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.status_code_retries = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.516706] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.status_code_retry_delay = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.516857] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.system_scope = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.517019] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.timeout = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.517180] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.trust_id = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.517341] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.user_domain_id = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.517508] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.user_domain_name = Default {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.517666] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.user_id = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.517860] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.username = nova {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.518097] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.518307] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] placement.version = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.518517] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] quota.cores = 20 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.518693] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] quota.count_usage_from_placement = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.518870] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.519079] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] quota.injected_file_content_bytes = 10240 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.519261] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] quota.injected_file_path_length = 255 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.519430] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] quota.injected_files = 5 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.519596] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] quota.instances = 10 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.519762] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] quota.key_pairs = 100 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.519926] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] quota.metadata_items = 128 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.520105] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] quota.ram = 51200 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.520307] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] quota.recheck_quota = True {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.520439] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] quota.server_group_members = 10 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.520604] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] quota.server_groups = 10 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.520785] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.520949] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.521129] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] scheduler.image_metadata_prefilter = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.521296] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.521459] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] scheduler.max_attempts = 3 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.521621] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] scheduler.max_placement_results = 1000 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.521787] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.521947] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] scheduler.query_placement_for_image_type_support = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.522122] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.522300] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] scheduler.workers = 2 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.522472] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.522642] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.522823] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.522993] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.523176] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.523341] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.523507] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.523700] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.523870] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] filter_scheduler.host_subset_size = 1 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.524055] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.524221] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.524387] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.524595] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] filter_scheduler.isolated_hosts = [] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.524715] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] filter_scheduler.isolated_images = [] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.524878] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.525050] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.525221] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.525382] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] filter_scheduler.pci_in_placement = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.525542] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.525701] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.525866] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.526059] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.526231] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.526395] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.526554] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] filter_scheduler.track_instance_changes = True {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.526732] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.526902] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] metrics.required = True {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.527079] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] metrics.weight_multiplier = 1.0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.527249] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.527413] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] metrics.weight_setting = [] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.527728] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.527977] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] serial_console.enabled = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.528120] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] serial_console.port_range = 10000:20000 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.528305] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.528476] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.528641] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] serial_console.serialproxy_port = 6083 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.528809] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] service_user.auth_section = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.528980] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] service_user.auth_type = password {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.529160] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] service_user.cafile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.529318] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] service_user.certfile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.529479] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] service_user.collect_timing = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.529640] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] service_user.insecure = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.529796] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] service_user.keyfile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.529964] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] service_user.send_service_user_token = True {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.530142] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] service_user.split_loggers = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.530315] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] service_user.timeout = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.530487] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] spice.agent_enabled = True {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.530650] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] spice.enabled = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.530965] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.531174] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.531346] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] spice.html5proxy_port = 6082 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.531507] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] spice.image_compression = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.531667] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] spice.jpeg_compression = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.531825] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] spice.playback_compression = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.531986] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] spice.require_secure = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.532170] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] spice.server_listen = 127.0.0.1 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.532338] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.532497] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] spice.streaming_mode = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.532654] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] spice.zlib_compression = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.532814] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] upgrade_levels.baseapi = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.532988] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] upgrade_levels.compute = auto {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.533164] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] upgrade_levels.conductor = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.533324] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] upgrade_levels.scheduler = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.533489] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vendordata_dynamic_auth.auth_section = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.533647] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vendordata_dynamic_auth.auth_type = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.533803] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vendordata_dynamic_auth.cafile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.533959] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vendordata_dynamic_auth.certfile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.534134] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.534297] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vendordata_dynamic_auth.insecure = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.534454] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vendordata_dynamic_auth.keyfile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.534612] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.534769] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vendordata_dynamic_auth.timeout = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.534938] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vmware.api_retry_count = 10 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.535109] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vmware.ca_file = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.535288] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vmware.cache_prefix = devstack-image-cache {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.535454] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vmware.cluster_name = testcl1 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.535617] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vmware.connection_pool_size = 10 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.535773] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vmware.console_delay_seconds = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.535991] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vmware.datastore_regex = ^datastore.* {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.536236] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.536414] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vmware.host_password = **** {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.536583] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vmware.host_port = 443 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.536757] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vmware.host_username = administrator@vsphere.local {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.536928] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vmware.insecure = True {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.537104] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vmware.integration_bridge = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.537273] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vmware.maximum_objects = 100 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.537431] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vmware.pbm_default_policy = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.537594] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vmware.pbm_enabled = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.537751] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vmware.pbm_wsdl_location = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.537919] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.538116] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vmware.serial_port_proxy_uri = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.538288] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vmware.serial_port_service_uri = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.538456] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vmware.task_poll_interval = 0.5 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.538627] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vmware.use_linked_clone = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.538796] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vmware.vnc_keymap = en-us {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.538961] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vmware.vnc_port = 5900 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.539140] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vmware.vnc_port_total = 10000 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.539345] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vnc.auth_schemes = ['none'] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.539571] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vnc.enabled = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.539861] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.540060] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.540240] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vnc.novncproxy_port = 6080 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.540418] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vnc.server_listen = 127.0.0.1 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.540588] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.540748] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vnc.vencrypt_ca_certs = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.540907] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vnc.vencrypt_client_cert = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.541075] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vnc.vencrypt_client_key = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.541261] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.541425] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] workarounds.disable_deep_image_inspection = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.541585] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.541745] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.541903] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.542076] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] workarounds.disable_rootwrap = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.542243] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] workarounds.enable_numa_live_migration = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.542403] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.542564] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.542724] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.542882] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] workarounds.libvirt_disable_apic = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.543052] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.543221] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.543382] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.543540] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.543702] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.543861] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.544030] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.544198] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.544359] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.544522] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.544707] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.544907] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] wsgi.client_socket_timeout = 900 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.545106] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] wsgi.default_pool_size = 1000 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.545281] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] wsgi.keep_alive = True {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.545448] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] wsgi.max_header_line = 16384 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.545611] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] wsgi.secure_proxy_ssl_header = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.545772] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] wsgi.ssl_ca_file = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.545934] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] wsgi.ssl_cert_file = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.546141] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] wsgi.ssl_key_file = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.546320] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] wsgi.tcp_keepidle = 600 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.546504] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.546673] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] zvm.ca_file = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.546832] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] zvm.cloud_connector_url = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.547133] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.547309] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] zvm.reachable_timeout = 300 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.547489] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_policy.enforce_new_defaults = True {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.547871] env[61868]: WARNING oslo_config.cfg [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 493.548100] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_policy.enforce_scope = True {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.548296] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_policy.policy_default_rule = default {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.548486] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.548666] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_policy.policy_file = policy.yaml {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.548844] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.549015] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.549189] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.549349] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.549514] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.549684] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.549859] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.550047] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] profiler.connection_string = messaging:// {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.550222] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] profiler.enabled = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.550391] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] profiler.es_doc_type = notification {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.550554] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] profiler.es_scroll_size = 10000 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.550720] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] profiler.es_scroll_time = 2m {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.550881] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] profiler.filter_error_trace = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.551062] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] profiler.hmac_keys = **** {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.551237] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] profiler.sentinel_service_name = mymaster {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.551403] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] profiler.socket_timeout = 0.1 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.551563] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] profiler.trace_requests = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.551721] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] profiler.trace_sqlalchemy = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.551901] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] profiler_jaeger.process_tags = {} {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.552073] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] profiler_jaeger.service_name_prefix = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.552241] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] profiler_otlp.service_name_prefix = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.552406] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] remote_debug.host = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.552566] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] remote_debug.port = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.552745] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.552909] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.553085] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.553252] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.553414] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.553572] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.553731] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.553891] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.554068] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.554452] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.554452] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.554568] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.554728] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.554898] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.555073] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.555244] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.555406] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.555622] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.555742] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.555921] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.556116] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.556287] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.556450] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.556617] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.556777] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.556939] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.557111] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.557279] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.557445] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.557610] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_rabbit.ssl = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.557783] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.557951] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.558161] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.558340] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.558510] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_rabbit.ssl_version = {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.558672] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.558860] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.559069] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_notifications.retry = -1 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.559280] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.559460] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_messaging_notifications.transport_url = **** {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.559633] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_limit.auth_section = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.559801] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_limit.auth_type = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.559961] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_limit.cafile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.560137] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_limit.certfile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.560302] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_limit.collect_timing = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.560469] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_limit.connect_retries = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.560615] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_limit.connect_retry_delay = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.560772] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_limit.endpoint_id = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.560931] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_limit.endpoint_override = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.561104] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_limit.insecure = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.561267] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_limit.keyfile = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.561424] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_limit.max_version = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.561579] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_limit.min_version = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.561734] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_limit.region_name = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.561894] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_limit.retriable_status_codes = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.562102] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_limit.service_name = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.562298] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_limit.service_type = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.562466] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_limit.split_loggers = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.562628] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_limit.status_code_retries = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.562788] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_limit.status_code_retry_delay = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.562946] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_limit.timeout = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.563122] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_limit.valid_interfaces = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.563284] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_limit.version = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.563450] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_reports.file_event_handler = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.563614] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.563773] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] oslo_reports.log_dir = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.563943] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.564119] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.564283] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.564452] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.564632] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.564789] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.564989] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.565128] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vif_plug_ovs_privileged.group = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.565288] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.565452] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.565613] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.565770] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] vif_plug_ovs_privileged.user = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.565964] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] os_vif_linux_bridge.flat_interface = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.566175] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.566354] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.566527] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.566701] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.566868] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.567048] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.567221] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.567401] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.567571] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] os_vif_ovs.isolate_vif = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.567738] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.567904] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.568114] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.568299] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] os_vif_ovs.ovsdb_interface = native {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.568464] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] os_vif_ovs.per_port_bridge = False {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.568637] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] privsep_osbrick.capabilities = [21] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.568797] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] privsep_osbrick.group = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.568956] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] privsep_osbrick.helper_command = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.569160] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.569328] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.569489] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] privsep_osbrick.user = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.569663] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.569824] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] nova_sys_admin.group = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.569983] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] nova_sys_admin.helper_command = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.570166] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.570330] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.570488] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] nova_sys_admin.user = None {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.570619] env[61868]: DEBUG oslo_service.service [None req-2f733d95-897a-4d6a-aa83-fd6520c94218 None None] ******************************************************************************** {{(pid=61868) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 493.571106] env[61868]: INFO nova.service [-] Starting compute node (version 0.0.1) [ 494.074404] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-034e479a-46a1-4775-a7fb-183405b5f705 None None] Getting list of instances from cluster (obj){ [ 494.074404] env[61868]: value = "domain-c8" [ 494.074404] env[61868]: _type = "ClusterComputeResource" [ 494.074404] env[61868]: } {{(pid=61868) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 494.075679] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddac43ec-d460-4440-a1f1-057811584c21 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 494.084910] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-034e479a-46a1-4775-a7fb-183405b5f705 None None] Got total of 0 instances {{(pid=61868) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 494.085459] env[61868]: WARNING nova.virt.vmwareapi.driver [None req-034e479a-46a1-4775-a7fb-183405b5f705 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 494.085933] env[61868]: INFO nova.virt.node [None req-034e479a-46a1-4775-a7fb-183405b5f705 None None] Generated node identity 83e67721-2ac9-4a23-aa31-82aca86979c8 [ 494.086221] env[61868]: INFO nova.virt.node [None req-034e479a-46a1-4775-a7fb-183405b5f705 None None] Wrote node identity 83e67721-2ac9-4a23-aa31-82aca86979c8 to /opt/stack/data/n-cpu-1/compute_id [ 494.589375] env[61868]: WARNING nova.compute.manager [None req-034e479a-46a1-4775-a7fb-183405b5f705 None None] Compute nodes ['83e67721-2ac9-4a23-aa31-82aca86979c8'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 495.596168] env[61868]: INFO nova.compute.manager [None req-034e479a-46a1-4775-a7fb-183405b5f705 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 496.600826] env[61868]: WARNING nova.compute.manager [None req-034e479a-46a1-4775-a7fb-183405b5f705 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 496.601197] env[61868]: DEBUG oslo_concurrency.lockutils [None req-034e479a-46a1-4775-a7fb-183405b5f705 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 496.601276] env[61868]: DEBUG oslo_concurrency.lockutils [None req-034e479a-46a1-4775-a7fb-183405b5f705 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 496.601433] env[61868]: DEBUG oslo_concurrency.lockutils [None req-034e479a-46a1-4775-a7fb-183405b5f705 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 496.601589] env[61868]: DEBUG nova.compute.resource_tracker [None req-034e479a-46a1-4775-a7fb-183405b5f705 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61868) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 496.602595] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f6ca600-1890-412c-ba4c-2573e82a5646 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 496.611156] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00735b4f-3e15-433d-957d-4a802d5ad161 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 496.624842] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-044b8ccd-0d28-4a4c-8a5d-820980a41ccc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 496.631230] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3431d05f-42fd-4dc0-ae96-140de9b9dc7e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 496.659628] env[61868]: DEBUG nova.compute.resource_tracker [None req-034e479a-46a1-4775-a7fb-183405b5f705 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181527MB free_disk=135GB free_vcpus=48 pci_devices=None {{(pid=61868) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 496.659768] env[61868]: DEBUG oslo_concurrency.lockutils [None req-034e479a-46a1-4775-a7fb-183405b5f705 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 496.659984] env[61868]: DEBUG oslo_concurrency.lockutils [None req-034e479a-46a1-4775-a7fb-183405b5f705 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 497.162303] env[61868]: WARNING nova.compute.resource_tracker [None req-034e479a-46a1-4775-a7fb-183405b5f705 None None] No compute node record for cpu-1:83e67721-2ac9-4a23-aa31-82aca86979c8: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 83e67721-2ac9-4a23-aa31-82aca86979c8 could not be found. [ 497.665904] env[61868]: INFO nova.compute.resource_tracker [None req-034e479a-46a1-4775-a7fb-183405b5f705 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 83e67721-2ac9-4a23-aa31-82aca86979c8 [ 499.174333] env[61868]: DEBUG nova.compute.resource_tracker [None req-034e479a-46a1-4775-a7fb-183405b5f705 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61868) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 499.174708] env[61868]: DEBUG nova.compute.resource_tracker [None req-034e479a-46a1-4775-a7fb-183405b5f705 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61868) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 499.321597] env[61868]: INFO nova.scheduler.client.report [None req-034e479a-46a1-4775-a7fb-183405b5f705 None None] [req-08b3bb4c-4862-4cc1-8612-c5220d970dbd] Created resource provider record via placement API for resource provider with UUID 83e67721-2ac9-4a23-aa31-82aca86979c8 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 499.337842] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb3c034e-65cf-451a-80d2-1f8e5aca333c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.345405] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26d4d65f-73ce-4f85-9d9a-0bcf05c4196b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.375355] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-def615e2-7959-49b9-abf1-6c9c4043048d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.381919] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a0d9efd-66f6-445d-b773-2dadd388cf7c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.394096] env[61868]: DEBUG nova.compute.provider_tree [None req-034e479a-46a1-4775-a7fb-183405b5f705 None None] Updating inventory in ProviderTree for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 499.958673] env[61868]: DEBUG nova.scheduler.client.report [None req-034e479a-46a1-4775-a7fb-183405b5f705 None None] Updated inventory for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 499.958911] env[61868]: DEBUG nova.compute.provider_tree [None req-034e479a-46a1-4775-a7fb-183405b5f705 None None] Updating resource provider 83e67721-2ac9-4a23-aa31-82aca86979c8 generation from 0 to 1 during operation: update_inventory {{(pid=61868) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 499.959108] env[61868]: DEBUG nova.compute.provider_tree [None req-034e479a-46a1-4775-a7fb-183405b5f705 None None] Updating inventory in ProviderTree for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 500.008779] env[61868]: DEBUG nova.compute.provider_tree [None req-034e479a-46a1-4775-a7fb-183405b5f705 None None] Updating resource provider 83e67721-2ac9-4a23-aa31-82aca86979c8 generation from 1 to 2 during operation: update_traits {{(pid=61868) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 500.513318] env[61868]: DEBUG nova.compute.resource_tracker [None req-034e479a-46a1-4775-a7fb-183405b5f705 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61868) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 500.513722] env[61868]: DEBUG oslo_concurrency.lockutils [None req-034e479a-46a1-4775-a7fb-183405b5f705 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.854s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 500.513722] env[61868]: DEBUG nova.service [None req-034e479a-46a1-4775-a7fb-183405b5f705 None None] Creating RPC server for service compute {{(pid=61868) start /opt/stack/nova/nova/service.py:186}} [ 500.526960] env[61868]: DEBUG nova.service [None req-034e479a-46a1-4775-a7fb-183405b5f705 None None] Join ServiceGroup membership for this service compute {{(pid=61868) start /opt/stack/nova/nova/service.py:203}} [ 500.527173] env[61868]: DEBUG nova.servicegroup.drivers.db [None req-034e479a-46a1-4775-a7fb-183405b5f705 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=61868) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 537.531067] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._sync_power_states {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 538.034201] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Getting list of instances from cluster (obj){ [ 538.034201] env[61868]: value = "domain-c8" [ 538.034201] env[61868]: _type = "ClusterComputeResource" [ 538.034201] env[61868]: } {{(pid=61868) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 538.036691] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33031d2d-f90c-4446-b2ce-481fef77b75b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.049337] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Got total of 0 instances {{(pid=61868) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 538.049896] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 538.052152] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Getting list of instances from cluster (obj){ [ 538.052152] env[61868]: value = "domain-c8" [ 538.052152] env[61868]: _type = "ClusterComputeResource" [ 538.052152] env[61868]: } {{(pid=61868) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 538.052152] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20be5a7c-bc9f-429d-86e6-895f26e42e65 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.060633] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Got total of 0 instances {{(pid=61868) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 540.217230] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] Acquiring lock "70adb078-edbe-4a66-b2c3-a6721bc3ffe6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.217528] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] Lock "70adb078-edbe-4a66-b2c3-a6721bc3ffe6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.720697] env[61868]: DEBUG nova.compute.manager [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 540.794234] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Acquiring lock "3442fbd9-83cf-4835-846c-0a32b2690598" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.794234] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Lock "3442fbd9-83cf-4835-846c-0a32b2690598" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.171833] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] Acquiring lock "38c58279-2caf-4d74-a73b-f3179e6f7f3d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.172067] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] Lock "38c58279-2caf-4d74-a73b-f3179e6f7f3d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.267965] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.268326] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.271274] env[61868]: INFO nova.compute.claims [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 541.297054] env[61868]: DEBUG nova.compute.manager [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 541.487250] env[61868]: DEBUG oslo_concurrency.lockutils [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "c8932fff-46b9-4d1d-ab0b-cc88823bd940" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.487665] env[61868]: DEBUG oslo_concurrency.lockutils [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "c8932fff-46b9-4d1d-ab0b-cc88823bd940" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.675799] env[61868]: DEBUG nova.compute.manager [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 541.825168] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.936450] env[61868]: DEBUG oslo_concurrency.lockutils [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] Acquiring lock "a22aee91-ad9b-4eb1-be66-315900bacf78" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.944498] env[61868]: DEBUG oslo_concurrency.lockutils [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] Lock "a22aee91-ad9b-4eb1-be66-315900bacf78" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.005s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.993613] env[61868]: DEBUG nova.compute.manager [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 542.218448] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.411593] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd0b77a3-04a3-430e-8261-b1e3c70d6fb2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.425621] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37ad70ce-24ba-4261-89ec-c9a79b7d008f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.460283] env[61868]: DEBUG nova.compute.manager [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 542.463928] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7216289-2ef7-40f2-bdd0-3fe32b44024d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.475445] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bffa9540-aba6-4c61-ba12-197ea7aa2015 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.497456] env[61868]: DEBUG nova.compute.provider_tree [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 542.522243] env[61868]: DEBUG oslo_concurrency.lockutils [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.842110] env[61868]: DEBUG oslo_concurrency.lockutils [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Acquiring lock "f7fd9e57-9bf1-47d8-b77e-c99436391d1c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.842439] env[61868]: DEBUG oslo_concurrency.lockutils [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Lock "f7fd9e57-9bf1-47d8-b77e-c99436391d1c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.984129] env[61868]: DEBUG oslo_concurrency.lockutils [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Acquiring lock "ddde9ee7-5198-42e5-86b9-f51727dfbf60" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.984441] env[61868]: DEBUG oslo_concurrency.lockutils [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Lock "ddde9ee7-5198-42e5-86b9-f51727dfbf60" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.991976] env[61868]: DEBUG oslo_concurrency.lockutils [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.002971] env[61868]: DEBUG nova.scheduler.client.report [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 543.347585] env[61868]: DEBUG nova.compute.manager [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 543.465857] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Acquiring lock "da571582-dfe5-434b-a947-37b075b31f97" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.466472] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Lock "da571582-dfe5-434b-a947-37b075b31f97" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.489319] env[61868]: DEBUG nova.compute.manager [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 543.510933] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.242s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 543.511554] env[61868]: DEBUG nova.compute.manager [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 543.518237] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.693s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.519328] env[61868]: INFO nova.compute.claims [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 543.885017] env[61868]: DEBUG oslo_concurrency.lockutils [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.968275] env[61868]: DEBUG nova.compute.manager [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: da571582-dfe5-434b-a947-37b075b31f97] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 544.010626] env[61868]: DEBUG oslo_concurrency.lockutils [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.029893] env[61868]: DEBUG nova.compute.utils [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 544.031842] env[61868]: DEBUG nova.compute.manager [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 544.032700] env[61868]: DEBUG nova.network.neutron [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 544.499236] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.526481] env[61868]: DEBUG nova.policy [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6cc79d056d4c4d4ea660bbdec518302e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5e073d46156b4205b7e3d80a0cb99a00', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 544.539914] env[61868]: DEBUG nova.compute.manager [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 544.693063] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-372645c0-b9cb-4b87-9673-4f5db855930a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.699663] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1adc9f8-f8ed-4ab8-9cd0-a00a0a385936 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.734096] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e9c011b-b59f-426e-8673-4afa8d7bb52a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.741123] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6293eef0-35fc-4f44-aa6a-b9dee807a5c0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.756099] env[61868]: DEBUG nova.compute.provider_tree [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 545.157900] env[61868]: DEBUG nova.network.neutron [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Successfully created port: 47de6438-6d4d-45eb-bf20-1cdce97a1da2 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 545.261309] env[61868]: DEBUG nova.scheduler.client.report [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 545.551115] env[61868]: DEBUG nova.compute.manager [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 545.572118] env[61868]: DEBUG nova.virt.hardware [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 545.572301] env[61868]: DEBUG nova.virt.hardware [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 545.572646] env[61868]: DEBUG nova.virt.hardware [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 545.572815] env[61868]: DEBUG nova.virt.hardware [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 545.572872] env[61868]: DEBUG nova.virt.hardware [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 545.572980] env[61868]: DEBUG nova.virt.hardware [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 545.573199] env[61868]: DEBUG nova.virt.hardware [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 545.573347] env[61868]: DEBUG nova.virt.hardware [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 545.573800] env[61868]: DEBUG nova.virt.hardware [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 545.574009] env[61868]: DEBUG nova.virt.hardware [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 545.574091] env[61868]: DEBUG nova.virt.hardware [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 545.574954] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c4d82ab-3215-42f0-8544-0e1d441f4610 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.584489] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f40d1e96-bacf-49d0-be94-fc76776da706 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.600675] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6b22b5a-0a37-4311-9c41-cf4336c76955 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.767929] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.250s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 545.769073] env[61868]: DEBUG nova.compute.manager [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 545.775176] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.553s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.778015] env[61868]: INFO nova.compute.claims [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 545.806637] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] Acquiring lock "be7273d1-963e-44bf-9862-35703e37081d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.806637] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] Lock "be7273d1-963e-44bf-9862-35703e37081d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.281510] env[61868]: DEBUG nova.compute.utils [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 546.289349] env[61868]: DEBUG nova.compute.manager [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 546.289614] env[61868]: DEBUG nova.network.neutron [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 546.310919] env[61868]: DEBUG nova.compute.manager [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] [instance: be7273d1-963e-44bf-9862-35703e37081d] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 546.482700] env[61868]: DEBUG nova.policy [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd2c48e56b59b4b78b3676b1261d63360', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ccc6a8fff76843038eaef30e154e511f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 546.789732] env[61868]: DEBUG nova.compute.manager [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 546.849718] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.963065] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90997cb9-0208-407a-85d5-e796746c47ef {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.972921] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53de33af-bfa8-436a-ad8b-861326e2f655 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.001083] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7370025-f7f9-40f0-977b-f262cd37b0e1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.008854] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96f009ce-f99c-4e4b-a574-7ccdb5536b47 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.022581] env[61868]: DEBUG nova.compute.provider_tree [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 547.525929] env[61868]: DEBUG nova.scheduler.client.report [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 547.801488] env[61868]: DEBUG nova.compute.manager [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 547.829522] env[61868]: DEBUG nova.virt.hardware [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 547.830769] env[61868]: DEBUG nova.virt.hardware [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 547.830769] env[61868]: DEBUG nova.virt.hardware [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 547.831063] env[61868]: DEBUG nova.virt.hardware [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 547.831185] env[61868]: DEBUG nova.virt.hardware [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 547.831332] env[61868]: DEBUG nova.virt.hardware [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 547.831780] env[61868]: DEBUG nova.virt.hardware [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 547.831780] env[61868]: DEBUG nova.virt.hardware [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 547.831877] env[61868]: DEBUG nova.virt.hardware [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 547.831985] env[61868]: DEBUG nova.virt.hardware [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 547.835143] env[61868]: DEBUG nova.virt.hardware [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 547.835143] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b3a2434-9e4f-4950-8879-84e3344c6f79 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.842462] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff45f451-29ca-4c48-b734-94f49fadd8ae {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.909033] env[61868]: DEBUG nova.network.neutron [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Successfully created port: d9a77d13-e9ba-4ee7-b899-0e84f7a76643 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 548.040014] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.269s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 548.040778] env[61868]: DEBUG nova.compute.manager [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 548.043933] env[61868]: DEBUG oslo_concurrency.lockutils [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.522s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.045536] env[61868]: INFO nova.compute.claims [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 548.551498] env[61868]: DEBUG nova.compute.utils [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 548.555308] env[61868]: DEBUG nova.compute.manager [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 548.555486] env[61868]: DEBUG nova.network.neutron [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 548.736548] env[61868]: DEBUG nova.policy [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6c45ed3d428f4e979adfaf1b02962de2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '133f4956fc394ec89057f6c44b9ff459', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 549.037695] env[61868]: ERROR nova.compute.manager [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 47de6438-6d4d-45eb-bf20-1cdce97a1da2, please check neutron logs for more information. [ 549.037695] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 549.037695] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 549.037695] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 549.037695] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 549.037695] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 549.037695] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 549.037695] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 549.037695] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 549.037695] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 549.037695] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 549.037695] env[61868]: ERROR nova.compute.manager raise self.value [ 549.037695] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 549.037695] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 549.037695] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 549.037695] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 549.038387] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 549.038387] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 549.038387] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 47de6438-6d4d-45eb-bf20-1cdce97a1da2, please check neutron logs for more information. [ 549.038387] env[61868]: ERROR nova.compute.manager [ 549.038387] env[61868]: Traceback (most recent call last): [ 549.038387] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 549.038387] env[61868]: listener.cb(fileno) [ 549.038387] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 549.038387] env[61868]: result = function(*args, **kwargs) [ 549.038387] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 549.038387] env[61868]: return func(*args, **kwargs) [ 549.038387] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 549.038387] env[61868]: raise e [ 549.038387] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 549.038387] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 549.038387] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 549.038387] env[61868]: created_port_ids = self._update_ports_for_instance( [ 549.038387] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 549.038387] env[61868]: with excutils.save_and_reraise_exception(): [ 549.038387] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 549.038387] env[61868]: self.force_reraise() [ 549.038387] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 549.038387] env[61868]: raise self.value [ 549.038387] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 549.038387] env[61868]: updated_port = self._update_port( [ 549.038387] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 549.038387] env[61868]: _ensure_no_port_binding_failure(port) [ 549.038387] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 549.038387] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 549.039178] env[61868]: nova.exception.PortBindingFailed: Binding failed for port 47de6438-6d4d-45eb-bf20-1cdce97a1da2, please check neutron logs for more information. [ 549.039178] env[61868]: Removing descriptor: 15 [ 549.040713] env[61868]: ERROR nova.compute.manager [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 47de6438-6d4d-45eb-bf20-1cdce97a1da2, please check neutron logs for more information. [ 549.040713] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Traceback (most recent call last): [ 549.040713] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 549.040713] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] yield resources [ 549.040713] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 549.040713] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] self.driver.spawn(context, instance, image_meta, [ 549.040713] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 549.040713] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 549.040713] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 549.040713] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] vm_ref = self.build_virtual_machine(instance, [ 549.040713] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 549.041088] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] vif_infos = vmwarevif.get_vif_info(self._session, [ 549.041088] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 549.041088] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] for vif in network_info: [ 549.041088] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 549.041088] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] return self._sync_wrapper(fn, *args, **kwargs) [ 549.041088] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 549.041088] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] self.wait() [ 549.041088] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 549.041088] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] self[:] = self._gt.wait() [ 549.041088] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 549.041088] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] return self._exit_event.wait() [ 549.041088] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 549.041088] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] result = hub.switch() [ 549.041434] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 549.041434] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] return self.greenlet.switch() [ 549.041434] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 549.041434] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] result = function(*args, **kwargs) [ 549.041434] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 549.041434] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] return func(*args, **kwargs) [ 549.041434] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 549.041434] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] raise e [ 549.041434] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 549.041434] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] nwinfo = self.network_api.allocate_for_instance( [ 549.041434] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 549.041434] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] created_port_ids = self._update_ports_for_instance( [ 549.041434] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 549.041770] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] with excutils.save_and_reraise_exception(): [ 549.041770] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 549.041770] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] self.force_reraise() [ 549.041770] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 549.041770] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] raise self.value [ 549.041770] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 549.041770] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] updated_port = self._update_port( [ 549.041770] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 549.041770] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] _ensure_no_port_binding_failure(port) [ 549.041770] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 549.041770] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] raise exception.PortBindingFailed(port_id=port['id']) [ 549.041770] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] nova.exception.PortBindingFailed: Binding failed for port 47de6438-6d4d-45eb-bf20-1cdce97a1da2, please check neutron logs for more information. [ 549.041770] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] [ 549.042260] env[61868]: INFO nova.compute.manager [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Terminating instance [ 549.043929] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] Acquiring lock "refresh_cache-70adb078-edbe-4a66-b2c3-a6721bc3ffe6" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 549.044114] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] Acquired lock "refresh_cache-70adb078-edbe-4a66-b2c3-a6721bc3ffe6" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 549.044268] env[61868]: DEBUG nova.network.neutron [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 549.053919] env[61868]: DEBUG nova.compute.manager [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 549.087125] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 549.087125] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 549.087222] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Starting heal instance info cache {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 549.087348] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Rebuilding the list of instances to heal {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 549.256026] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bed42182-511e-4ee8-baf4-c1009abf2778 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.264580] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15332739-3b36-48b2-8665-92e30265a284 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.298428] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a96c8ab7-5852-44d3-8113-11fd559e2949 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.306256] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-120687cd-1793-46f7-8e79-85680ec9a2b9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.320834] env[61868]: DEBUG nova.compute.provider_tree [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 549.592652] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Skipping network cache update for instance because it is Building. {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 549.592652] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Skipping network cache update for instance because it is Building. {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 549.592652] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Skipping network cache update for instance because it is Building. {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 549.592652] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Skipping network cache update for instance because it is Building. {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 549.592652] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Didn't find any instances for network info cache update. {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 549.592652] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 549.593050] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 549.593840] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 549.595506] env[61868]: DEBUG nova.network.neutron [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 549.602852] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 549.602852] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 549.603751] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 549.603997] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61868) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 549.605025] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager.update_available_resource {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 549.660338] env[61868]: DEBUG nova.network.neutron [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Successfully created port: 720aed11-8a04-47a0-85cb-a0bfe9c615c1 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 549.824657] env[61868]: DEBUG nova.scheduler.client.report [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 549.925122] env[61868]: DEBUG nova.network.neutron [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.067756] env[61868]: DEBUG nova.compute.manager [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 550.111143] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.115115] env[61868]: DEBUG nova.virt.hardware [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 550.115115] env[61868]: DEBUG nova.virt.hardware [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 550.115115] env[61868]: DEBUG nova.virt.hardware [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 550.115304] env[61868]: DEBUG nova.virt.hardware [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 550.115304] env[61868]: DEBUG nova.virt.hardware [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 550.115304] env[61868]: DEBUG nova.virt.hardware [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 550.119016] env[61868]: DEBUG nova.virt.hardware [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 550.119016] env[61868]: DEBUG nova.virt.hardware [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 550.119016] env[61868]: DEBUG nova.virt.hardware [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 550.119016] env[61868]: DEBUG nova.virt.hardware [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 550.119016] env[61868]: DEBUG nova.virt.hardware [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 550.119237] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1627f82e-88cc-4cba-91d0-a6f34d96d7d7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.127076] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7056b9f1-fd44-4639-b42d-90d4f61005ab {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.329658] env[61868]: DEBUG oslo_concurrency.lockutils [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.285s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 550.330250] env[61868]: DEBUG nova.compute.manager [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 550.334412] env[61868]: DEBUG oslo_concurrency.lockutils [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.342s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.335915] env[61868]: INFO nova.compute.claims [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 550.428425] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] Releasing lock "refresh_cache-70adb078-edbe-4a66-b2c3-a6721bc3ffe6" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 550.428878] env[61868]: DEBUG nova.compute.manager [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 550.429061] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 550.429624] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0589d642-28b8-4949-824f-258d344015a0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.439053] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64cfa5e6-a0fa-43fc-b7e7-6b8e9bd368b0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.468199] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 70adb078-edbe-4a66-b2c3-a6721bc3ffe6 could not be found. [ 550.468199] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 550.468199] env[61868]: INFO nova.compute.manager [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Took 0.04 seconds to destroy the instance on the hypervisor. [ 550.468199] env[61868]: DEBUG oslo.service.loopingcall [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 550.468199] env[61868]: DEBUG nova.compute.manager [-] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 550.468392] env[61868]: DEBUG nova.network.neutron [-] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 550.506345] env[61868]: DEBUG nova.compute.manager [req-31f4bb0b-4f82-4e62-9291-ec48f34856f2 req-93b80521-a6cc-4b52-b7b5-4f2f87aaece8 service nova] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Received event network-changed-47de6438-6d4d-45eb-bf20-1cdce97a1da2 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 550.506345] env[61868]: DEBUG nova.compute.manager [req-31f4bb0b-4f82-4e62-9291-ec48f34856f2 req-93b80521-a6cc-4b52-b7b5-4f2f87aaece8 service nova] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Refreshing instance network info cache due to event network-changed-47de6438-6d4d-45eb-bf20-1cdce97a1da2. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 550.506345] env[61868]: DEBUG oslo_concurrency.lockutils [req-31f4bb0b-4f82-4e62-9291-ec48f34856f2 req-93b80521-a6cc-4b52-b7b5-4f2f87aaece8 service nova] Acquiring lock "refresh_cache-70adb078-edbe-4a66-b2c3-a6721bc3ffe6" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 550.506345] env[61868]: DEBUG oslo_concurrency.lockutils [req-31f4bb0b-4f82-4e62-9291-ec48f34856f2 req-93b80521-a6cc-4b52-b7b5-4f2f87aaece8 service nova] Acquired lock "refresh_cache-70adb078-edbe-4a66-b2c3-a6721bc3ffe6" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 550.506345] env[61868]: DEBUG nova.network.neutron [req-31f4bb0b-4f82-4e62-9291-ec48f34856f2 req-93b80521-a6cc-4b52-b7b5-4f2f87aaece8 service nova] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Refreshing network info cache for port 47de6438-6d4d-45eb-bf20-1cdce97a1da2 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 550.533732] env[61868]: DEBUG nova.network.neutron [-] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 550.638249] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] Acquiring lock "7965772b-4dd7-4807-8be4-67c6565f9304" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.638943] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] Lock "7965772b-4dd7-4807-8be4-67c6565f9304" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.845553] env[61868]: DEBUG nova.compute.utils [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 550.849431] env[61868]: DEBUG nova.compute.manager [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 550.849537] env[61868]: DEBUG nova.network.neutron [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 550.957080] env[61868]: DEBUG nova.policy [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fd6ed09d33f6491b83dc72cef31bfd29', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '812e115cc1114d0e8536924c4f5556f8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 551.037385] env[61868]: DEBUG nova.network.neutron [-] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.061039] env[61868]: DEBUG nova.network.neutron [req-31f4bb0b-4f82-4e62-9291-ec48f34856f2 req-93b80521-a6cc-4b52-b7b5-4f2f87aaece8 service nova] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 551.145127] env[61868]: DEBUG nova.compute.manager [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 551.350187] env[61868]: DEBUG nova.compute.manager [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 551.483070] env[61868]: DEBUG nova.network.neutron [req-31f4bb0b-4f82-4e62-9291-ec48f34856f2 req-93b80521-a6cc-4b52-b7b5-4f2f87aaece8 service nova] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.540565] env[61868]: INFO nova.compute.manager [-] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Took 1.07 seconds to deallocate network for instance. [ 551.547687] env[61868]: DEBUG nova.compute.claims [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 551.547783] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.632560] env[61868]: ERROR nova.compute.manager [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d9a77d13-e9ba-4ee7-b899-0e84f7a76643, please check neutron logs for more information. [ 551.632560] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 551.632560] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 551.632560] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 551.632560] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 551.632560] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 551.632560] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 551.632560] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 551.632560] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.632560] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 551.632560] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.632560] env[61868]: ERROR nova.compute.manager raise self.value [ 551.632560] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 551.632560] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 551.632560] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.632560] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 551.633124] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.633124] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 551.633124] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d9a77d13-e9ba-4ee7-b899-0e84f7a76643, please check neutron logs for more information. [ 551.633124] env[61868]: ERROR nova.compute.manager [ 551.633124] env[61868]: Traceback (most recent call last): [ 551.633124] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 551.633124] env[61868]: listener.cb(fileno) [ 551.633124] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 551.633124] env[61868]: result = function(*args, **kwargs) [ 551.633124] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 551.633124] env[61868]: return func(*args, **kwargs) [ 551.633124] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 551.633124] env[61868]: raise e [ 551.633124] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 551.633124] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 551.633124] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 551.633124] env[61868]: created_port_ids = self._update_ports_for_instance( [ 551.633124] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 551.633124] env[61868]: with excutils.save_and_reraise_exception(): [ 551.633124] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.633124] env[61868]: self.force_reraise() [ 551.633124] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.633124] env[61868]: raise self.value [ 551.633124] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 551.633124] env[61868]: updated_port = self._update_port( [ 551.633124] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.633124] env[61868]: _ensure_no_port_binding_failure(port) [ 551.633124] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.633124] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 551.634842] env[61868]: nova.exception.PortBindingFailed: Binding failed for port d9a77d13-e9ba-4ee7-b899-0e84f7a76643, please check neutron logs for more information. [ 551.634842] env[61868]: Removing descriptor: 16 [ 551.634842] env[61868]: ERROR nova.compute.manager [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d9a77d13-e9ba-4ee7-b899-0e84f7a76643, please check neutron logs for more information. [ 551.634842] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Traceback (most recent call last): [ 551.634842] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 551.634842] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] yield resources [ 551.634842] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 551.634842] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] self.driver.spawn(context, instance, image_meta, [ 551.634842] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 551.634842] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] self._vmops.spawn(context, instance, image_meta, injected_files, [ 551.634842] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 551.634842] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] vm_ref = self.build_virtual_machine(instance, [ 551.635422] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 551.635422] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] vif_infos = vmwarevif.get_vif_info(self._session, [ 551.635422] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 551.635422] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] for vif in network_info: [ 551.635422] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 551.635422] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] return self._sync_wrapper(fn, *args, **kwargs) [ 551.635422] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 551.635422] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] self.wait() [ 551.635422] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 551.635422] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] self[:] = self._gt.wait() [ 551.635422] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 551.635422] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] return self._exit_event.wait() [ 551.635422] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 551.635899] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] result = hub.switch() [ 551.635899] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 551.635899] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] return self.greenlet.switch() [ 551.635899] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 551.635899] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] result = function(*args, **kwargs) [ 551.635899] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 551.635899] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] return func(*args, **kwargs) [ 551.635899] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 551.635899] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] raise e [ 551.635899] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 551.635899] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] nwinfo = self.network_api.allocate_for_instance( [ 551.635899] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 551.635899] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] created_port_ids = self._update_ports_for_instance( [ 551.636462] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 551.636462] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] with excutils.save_and_reraise_exception(): [ 551.636462] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.636462] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] self.force_reraise() [ 551.636462] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.636462] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] raise self.value [ 551.636462] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 551.636462] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] updated_port = self._update_port( [ 551.636462] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.636462] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] _ensure_no_port_binding_failure(port) [ 551.636462] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.636462] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] raise exception.PortBindingFailed(port_id=port['id']) [ 551.638038] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] nova.exception.PortBindingFailed: Binding failed for port d9a77d13-e9ba-4ee7-b899-0e84f7a76643, please check neutron logs for more information. [ 551.638038] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] [ 551.638038] env[61868]: INFO nova.compute.manager [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Terminating instance [ 551.641058] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ea734cd-5bba-4efb-bf3d-4690bc6356bb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.644022] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Acquiring lock "refresh_cache-3442fbd9-83cf-4835-846c-0a32b2690598" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 551.644201] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Acquired lock "refresh_cache-3442fbd9-83cf-4835-846c-0a32b2690598" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 551.644338] env[61868]: DEBUG nova.network.neutron [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 551.653957] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1467ace-87a9-4514-9f2c-e1d2d698c9fb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.696547] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afade355-092f-4083-b353-cd6a8e588584 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.704557] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6827bce8-0649-4397-9733-69d4207ecbdc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.722139] env[61868]: DEBUG nova.compute.provider_tree [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 551.723937] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.928899] env[61868]: DEBUG nova.network.neutron [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Successfully created port: da55d5ca-ef4c-407e-81d9-457f3207cc1e {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 551.987101] env[61868]: DEBUG oslo_concurrency.lockutils [req-31f4bb0b-4f82-4e62-9291-ec48f34856f2 req-93b80521-a6cc-4b52-b7b5-4f2f87aaece8 service nova] Releasing lock "refresh_cache-70adb078-edbe-4a66-b2c3-a6721bc3ffe6" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 552.177335] env[61868]: DEBUG nova.network.neutron [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 552.229567] env[61868]: DEBUG nova.scheduler.client.report [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 552.235125] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Acquiring lock "a499452b-9d48-4022-8d4c-0c66c6dde9ba" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.235609] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Lock "a499452b-9d48-4022-8d4c-0c66c6dde9ba" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.364171] env[61868]: DEBUG nova.compute.manager [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 552.397119] env[61868]: DEBUG nova.virt.hardware [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 552.397408] env[61868]: DEBUG nova.virt.hardware [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 552.397577] env[61868]: DEBUG nova.virt.hardware [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 552.397814] env[61868]: DEBUG nova.virt.hardware [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 552.397904] env[61868]: DEBUG nova.virt.hardware [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 552.398047] env[61868]: DEBUG nova.virt.hardware [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 552.398289] env[61868]: DEBUG nova.virt.hardware [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 552.398494] env[61868]: DEBUG nova.virt.hardware [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 552.398568] env[61868]: DEBUG nova.virt.hardware [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 552.398813] env[61868]: DEBUG nova.virt.hardware [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 552.399009] env[61868]: DEBUG nova.virt.hardware [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 552.399996] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-503fd321-fee6-46d2-b440-8f68194cf369 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.408423] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81bf5b9b-2d8e-4fe9-b069-91672099780c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.449611] env[61868]: DEBUG nova.network.neutron [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.736322] env[61868]: DEBUG oslo_concurrency.lockutils [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.402s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 552.736913] env[61868]: DEBUG nova.compute.manager [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 552.741273] env[61868]: DEBUG oslo_concurrency.lockutils [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.857s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.742552] env[61868]: INFO nova.compute.claims [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 552.955685] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Releasing lock "refresh_cache-3442fbd9-83cf-4835-846c-0a32b2690598" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 552.955685] env[61868]: DEBUG nova.compute.manager [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 552.955685] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 552.955685] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a19ada38-2ace-4057-bf06-8e602a9b65f5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.965106] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01e1584d-0bb6-4ce2-ae73-2953bc629976 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.990365] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3442fbd9-83cf-4835-846c-0a32b2690598 could not be found. [ 552.990595] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 552.990795] env[61868]: INFO nova.compute.manager [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Took 0.04 seconds to destroy the instance on the hypervisor. [ 552.991100] env[61868]: DEBUG oslo.service.loopingcall [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 552.991257] env[61868]: DEBUG nova.compute.manager [-] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 552.991348] env[61868]: DEBUG nova.network.neutron [-] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 553.022728] env[61868]: DEBUG nova.network.neutron [-] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 553.247141] env[61868]: DEBUG nova.compute.utils [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 553.250163] env[61868]: DEBUG nova.compute.manager [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 553.250334] env[61868]: DEBUG nova.network.neutron [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 553.274826] env[61868]: ERROR nova.compute.manager [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 720aed11-8a04-47a0-85cb-a0bfe9c615c1, please check neutron logs for more information. [ 553.274826] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 553.274826] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 553.274826] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 553.274826] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 553.274826] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 553.274826] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 553.274826] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 553.274826] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 553.274826] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 553.274826] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 553.274826] env[61868]: ERROR nova.compute.manager raise self.value [ 553.274826] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 553.274826] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 553.274826] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 553.274826] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 553.275302] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 553.275302] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 553.275302] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 720aed11-8a04-47a0-85cb-a0bfe9c615c1, please check neutron logs for more information. [ 553.275302] env[61868]: ERROR nova.compute.manager [ 553.275302] env[61868]: Traceback (most recent call last): [ 553.275302] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 553.275302] env[61868]: listener.cb(fileno) [ 553.275302] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 553.275302] env[61868]: result = function(*args, **kwargs) [ 553.275302] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 553.275302] env[61868]: return func(*args, **kwargs) [ 553.275302] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 553.275302] env[61868]: raise e [ 553.275302] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 553.275302] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 553.275302] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 553.275302] env[61868]: created_port_ids = self._update_ports_for_instance( [ 553.275302] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 553.275302] env[61868]: with excutils.save_and_reraise_exception(): [ 553.275302] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 553.275302] env[61868]: self.force_reraise() [ 553.275302] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 553.275302] env[61868]: raise self.value [ 553.275302] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 553.275302] env[61868]: updated_port = self._update_port( [ 553.275302] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 553.275302] env[61868]: _ensure_no_port_binding_failure(port) [ 553.275302] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 553.275302] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 553.275978] env[61868]: nova.exception.PortBindingFailed: Binding failed for port 720aed11-8a04-47a0-85cb-a0bfe9c615c1, please check neutron logs for more information. [ 553.275978] env[61868]: Removing descriptor: 17 [ 553.276531] env[61868]: ERROR nova.compute.manager [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 720aed11-8a04-47a0-85cb-a0bfe9c615c1, please check neutron logs for more information. [ 553.276531] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Traceback (most recent call last): [ 553.276531] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 553.276531] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] yield resources [ 553.276531] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 553.276531] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] self.driver.spawn(context, instance, image_meta, [ 553.276531] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 553.276531] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 553.276531] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 553.276531] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] vm_ref = self.build_virtual_machine(instance, [ 553.276531] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 553.276881] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] vif_infos = vmwarevif.get_vif_info(self._session, [ 553.276881] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 553.276881] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] for vif in network_info: [ 553.276881] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 553.276881] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] return self._sync_wrapper(fn, *args, **kwargs) [ 553.276881] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 553.276881] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] self.wait() [ 553.276881] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 553.276881] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] self[:] = self._gt.wait() [ 553.276881] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 553.276881] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] return self._exit_event.wait() [ 553.276881] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 553.276881] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] result = hub.switch() [ 553.277250] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 553.277250] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] return self.greenlet.switch() [ 553.277250] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 553.277250] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] result = function(*args, **kwargs) [ 553.277250] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 553.277250] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] return func(*args, **kwargs) [ 553.277250] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 553.277250] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] raise e [ 553.277250] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 553.277250] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] nwinfo = self.network_api.allocate_for_instance( [ 553.277250] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 553.277250] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] created_port_ids = self._update_ports_for_instance( [ 553.277250] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 553.277569] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] with excutils.save_and_reraise_exception(): [ 553.277569] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 553.277569] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] self.force_reraise() [ 553.277569] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 553.277569] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] raise self.value [ 553.277569] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 553.277569] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] updated_port = self._update_port( [ 553.277569] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 553.277569] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] _ensure_no_port_binding_failure(port) [ 553.277569] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 553.277569] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] raise exception.PortBindingFailed(port_id=port['id']) [ 553.277569] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] nova.exception.PortBindingFailed: Binding failed for port 720aed11-8a04-47a0-85cb-a0bfe9c615c1, please check neutron logs for more information. [ 553.277569] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] [ 553.277915] env[61868]: INFO nova.compute.manager [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Terminating instance [ 553.281800] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] Acquiring lock "refresh_cache-38c58279-2caf-4d74-a73b-f3179e6f7f3d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 553.281800] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] Acquired lock "refresh_cache-38c58279-2caf-4d74-a73b-f3179e6f7f3d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 553.281800] env[61868]: DEBUG nova.network.neutron [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 553.356728] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] Acquiring lock "5e812408-6f27-4b52-9564-4063f4526340" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.356728] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] Lock "5e812408-6f27-4b52-9564-4063f4526340" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.452025] env[61868]: DEBUG nova.policy [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1c545651eeca41a58d1a4c696d48ec48', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5fc8b84aac2841418bd8e08d4fd6a825', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 553.525373] env[61868]: DEBUG nova.network.neutron [-] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.751060] env[61868]: DEBUG nova.compute.manager [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 553.854165] env[61868]: DEBUG nova.network.neutron [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 553.944470] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bd4e9e9-1121-42bd-9606-51264a71d320 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.954151] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f846f614-d67c-4d95-b64c-a6ea1678d4b6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.993674] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c209c64b-295e-4a8d-91b8-05b1557c81c2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.001969] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4277aa0e-2294-49bd-91b5-c78da592bb3d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.018104] env[61868]: DEBUG nova.compute.provider_tree [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 554.029468] env[61868]: INFO nova.compute.manager [-] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Took 1.04 seconds to deallocate network for instance. [ 554.031832] env[61868]: DEBUG nova.compute.claims [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 554.032023] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.101837] env[61868]: DEBUG nova.network.neutron [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 554.165017] env[61868]: DEBUG nova.compute.manager [req-a9ba8edf-b8ee-405a-9dcc-d90345fdf7bc req-76981ad8-3735-46f5-9a06-de4bf61a42bb service nova] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Received event network-changed-720aed11-8a04-47a0-85cb-a0bfe9c615c1 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 554.165017] env[61868]: DEBUG nova.compute.manager [req-a9ba8edf-b8ee-405a-9dcc-d90345fdf7bc req-76981ad8-3735-46f5-9a06-de4bf61a42bb service nova] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Refreshing instance network info cache due to event network-changed-720aed11-8a04-47a0-85cb-a0bfe9c615c1. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 554.165017] env[61868]: DEBUG oslo_concurrency.lockutils [req-a9ba8edf-b8ee-405a-9dcc-d90345fdf7bc req-76981ad8-3735-46f5-9a06-de4bf61a42bb service nova] Acquiring lock "refresh_cache-38c58279-2caf-4d74-a73b-f3179e6f7f3d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 554.187825] env[61868]: DEBUG nova.compute.manager [req-1900d0fb-f5d7-4023-b4b4-6aa2a789f668 req-2ab76447-eef3-4df1-ba01-6f36e6a2d7cf service nova] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Received event network-vif-deleted-47de6438-6d4d-45eb-bf20-1cdce97a1da2 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 554.188659] env[61868]: DEBUG nova.compute.manager [req-1900d0fb-f5d7-4023-b4b4-6aa2a789f668 req-2ab76447-eef3-4df1-ba01-6f36e6a2d7cf service nova] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Received event network-changed-d9a77d13-e9ba-4ee7-b899-0e84f7a76643 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 554.188659] env[61868]: DEBUG nova.compute.manager [req-1900d0fb-f5d7-4023-b4b4-6aa2a789f668 req-2ab76447-eef3-4df1-ba01-6f36e6a2d7cf service nova] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Refreshing instance network info cache due to event network-changed-d9a77d13-e9ba-4ee7-b899-0e84f7a76643. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 554.188659] env[61868]: DEBUG oslo_concurrency.lockutils [req-1900d0fb-f5d7-4023-b4b4-6aa2a789f668 req-2ab76447-eef3-4df1-ba01-6f36e6a2d7cf service nova] Acquiring lock "refresh_cache-3442fbd9-83cf-4835-846c-0a32b2690598" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 554.188659] env[61868]: DEBUG oslo_concurrency.lockutils [req-1900d0fb-f5d7-4023-b4b4-6aa2a789f668 req-2ab76447-eef3-4df1-ba01-6f36e6a2d7cf service nova] Acquired lock "refresh_cache-3442fbd9-83cf-4835-846c-0a32b2690598" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 554.188873] env[61868]: DEBUG nova.network.neutron [req-1900d0fb-f5d7-4023-b4b4-6aa2a789f668 req-2ab76447-eef3-4df1-ba01-6f36e6a2d7cf service nova] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Refreshing network info cache for port d9a77d13-e9ba-4ee7-b899-0e84f7a76643 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 554.522152] env[61868]: DEBUG nova.scheduler.client.report [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 554.605579] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] Releasing lock "refresh_cache-38c58279-2caf-4d74-a73b-f3179e6f7f3d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 554.605579] env[61868]: DEBUG nova.compute.manager [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 554.605579] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 554.606766] env[61868]: DEBUG oslo_concurrency.lockutils [req-a9ba8edf-b8ee-405a-9dcc-d90345fdf7bc req-76981ad8-3735-46f5-9a06-de4bf61a42bb service nova] Acquired lock "refresh_cache-38c58279-2caf-4d74-a73b-f3179e6f7f3d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 554.606894] env[61868]: DEBUG nova.network.neutron [req-a9ba8edf-b8ee-405a-9dcc-d90345fdf7bc req-76981ad8-3735-46f5-9a06-de4bf61a42bb service nova] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Refreshing network info cache for port 720aed11-8a04-47a0-85cb-a0bfe9c615c1 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 554.612462] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4046aad5-af52-4359-a633-5ae5263a076b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.624510] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7911aa19-7ec6-483e-b9b1-4f427027688e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.649649] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 38c58279-2caf-4d74-a73b-f3179e6f7f3d could not be found. [ 554.649901] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 554.650091] env[61868]: INFO nova.compute.manager [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Took 0.04 seconds to destroy the instance on the hypervisor. [ 554.650336] env[61868]: DEBUG oslo.service.loopingcall [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 554.650532] env[61868]: DEBUG nova.compute.manager [-] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 554.650623] env[61868]: DEBUG nova.network.neutron [-] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 554.702967] env[61868]: DEBUG nova.network.neutron [-] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 554.740029] env[61868]: DEBUG nova.network.neutron [req-1900d0fb-f5d7-4023-b4b4-6aa2a789f668 req-2ab76447-eef3-4df1-ba01-6f36e6a2d7cf service nova] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 554.742585] env[61868]: DEBUG nova.network.neutron [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Successfully created port: 286e88a8-7583-40d1-833c-9c34c435ca30 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 554.766394] env[61868]: DEBUG nova.compute.manager [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 554.794931] env[61868]: DEBUG nova.virt.hardware [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 554.795212] env[61868]: DEBUG nova.virt.hardware [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 554.795570] env[61868]: DEBUG nova.virt.hardware [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 554.795647] env[61868]: DEBUG nova.virt.hardware [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 554.796143] env[61868]: DEBUG nova.virt.hardware [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 554.796308] env[61868]: DEBUG nova.virt.hardware [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 554.796516] env[61868]: DEBUG nova.virt.hardware [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 554.796674] env[61868]: DEBUG nova.virt.hardware [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 554.796837] env[61868]: DEBUG nova.virt.hardware [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 554.796997] env[61868]: DEBUG nova.virt.hardware [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 554.797772] env[61868]: DEBUG nova.virt.hardware [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 554.798524] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93238e1e-91fe-457d-9ff3-6dd19ff38000 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.807447] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9e635e6-eadf-42bb-a4ab-2f37a306c02e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.951627] env[61868]: DEBUG nova.network.neutron [req-1900d0fb-f5d7-4023-b4b4-6aa2a789f668 req-2ab76447-eef3-4df1-ba01-6f36e6a2d7cf service nova] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 555.030860] env[61868]: DEBUG oslo_concurrency.lockutils [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.290s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 555.031409] env[61868]: DEBUG nova.compute.manager [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 555.034153] env[61868]: DEBUG oslo_concurrency.lockutils [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.024s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.038098] env[61868]: INFO nova.compute.claims [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 555.054100] env[61868]: ERROR nova.compute.manager [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port da55d5ca-ef4c-407e-81d9-457f3207cc1e, please check neutron logs for more information. [ 555.054100] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 555.054100] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 555.054100] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 555.054100] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 555.054100] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 555.054100] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 555.054100] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 555.054100] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 555.054100] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 555.054100] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 555.054100] env[61868]: ERROR nova.compute.manager raise self.value [ 555.054100] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 555.054100] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 555.054100] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 555.054100] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 555.054943] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 555.054943] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 555.054943] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port da55d5ca-ef4c-407e-81d9-457f3207cc1e, please check neutron logs for more information. [ 555.054943] env[61868]: ERROR nova.compute.manager [ 555.054943] env[61868]: Traceback (most recent call last): [ 555.054943] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 555.054943] env[61868]: listener.cb(fileno) [ 555.054943] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 555.054943] env[61868]: result = function(*args, **kwargs) [ 555.054943] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 555.054943] env[61868]: return func(*args, **kwargs) [ 555.054943] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 555.054943] env[61868]: raise e [ 555.054943] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 555.054943] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 555.054943] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 555.054943] env[61868]: created_port_ids = self._update_ports_for_instance( [ 555.054943] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 555.054943] env[61868]: with excutils.save_and_reraise_exception(): [ 555.054943] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 555.054943] env[61868]: self.force_reraise() [ 555.054943] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 555.054943] env[61868]: raise self.value [ 555.054943] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 555.054943] env[61868]: updated_port = self._update_port( [ 555.054943] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 555.054943] env[61868]: _ensure_no_port_binding_failure(port) [ 555.054943] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 555.054943] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 555.055646] env[61868]: nova.exception.PortBindingFailed: Binding failed for port da55d5ca-ef4c-407e-81d9-457f3207cc1e, please check neutron logs for more information. [ 555.055646] env[61868]: Removing descriptor: 15 [ 555.055646] env[61868]: ERROR nova.compute.manager [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port da55d5ca-ef4c-407e-81d9-457f3207cc1e, please check neutron logs for more information. [ 555.055646] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Traceback (most recent call last): [ 555.055646] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 555.055646] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] yield resources [ 555.055646] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 555.055646] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] self.driver.spawn(context, instance, image_meta, [ 555.055646] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 555.055646] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] self._vmops.spawn(context, instance, image_meta, injected_files, [ 555.055646] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 555.055646] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] vm_ref = self.build_virtual_machine(instance, [ 555.055943] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 555.055943] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] vif_infos = vmwarevif.get_vif_info(self._session, [ 555.055943] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 555.055943] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] for vif in network_info: [ 555.055943] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 555.055943] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] return self._sync_wrapper(fn, *args, **kwargs) [ 555.055943] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 555.055943] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] self.wait() [ 555.055943] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 555.055943] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] self[:] = self._gt.wait() [ 555.055943] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 555.055943] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] return self._exit_event.wait() [ 555.055943] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 555.056311] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] result = hub.switch() [ 555.056311] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 555.056311] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] return self.greenlet.switch() [ 555.056311] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 555.056311] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] result = function(*args, **kwargs) [ 555.056311] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 555.056311] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] return func(*args, **kwargs) [ 555.056311] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 555.056311] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] raise e [ 555.056311] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 555.056311] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] nwinfo = self.network_api.allocate_for_instance( [ 555.056311] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 555.056311] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] created_port_ids = self._update_ports_for_instance( [ 555.056697] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 555.056697] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] with excutils.save_and_reraise_exception(): [ 555.056697] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 555.056697] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] self.force_reraise() [ 555.056697] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 555.056697] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] raise self.value [ 555.056697] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 555.056697] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] updated_port = self._update_port( [ 555.056697] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 555.056697] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] _ensure_no_port_binding_failure(port) [ 555.056697] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 555.056697] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] raise exception.PortBindingFailed(port_id=port['id']) [ 555.057016] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] nova.exception.PortBindingFailed: Binding failed for port da55d5ca-ef4c-407e-81d9-457f3207cc1e, please check neutron logs for more information. [ 555.057016] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] [ 555.057016] env[61868]: INFO nova.compute.manager [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Terminating instance [ 555.057016] env[61868]: DEBUG oslo_concurrency.lockutils [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "refresh_cache-c8932fff-46b9-4d1d-ab0b-cc88823bd940" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 555.057016] env[61868]: DEBUG oslo_concurrency.lockutils [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquired lock "refresh_cache-c8932fff-46b9-4d1d-ab0b-cc88823bd940" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 555.057016] env[61868]: DEBUG nova.network.neutron [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 555.168055] env[61868]: DEBUG nova.network.neutron [req-a9ba8edf-b8ee-405a-9dcc-d90345fdf7bc req-76981ad8-3735-46f5-9a06-de4bf61a42bb service nova] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 555.208717] env[61868]: DEBUG nova.network.neutron [-] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 555.382714] env[61868]: DEBUG nova.network.neutron [req-a9ba8edf-b8ee-405a-9dcc-d90345fdf7bc req-76981ad8-3735-46f5-9a06-de4bf61a42bb service nova] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 555.458318] env[61868]: DEBUG oslo_concurrency.lockutils [req-1900d0fb-f5d7-4023-b4b4-6aa2a789f668 req-2ab76447-eef3-4df1-ba01-6f36e6a2d7cf service nova] Releasing lock "refresh_cache-3442fbd9-83cf-4835-846c-0a32b2690598" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 555.458589] env[61868]: DEBUG nova.compute.manager [req-1900d0fb-f5d7-4023-b4b4-6aa2a789f668 req-2ab76447-eef3-4df1-ba01-6f36e6a2d7cf service nova] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Received event network-vif-deleted-d9a77d13-e9ba-4ee7-b899-0e84f7a76643 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 555.543203] env[61868]: DEBUG nova.compute.utils [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 555.550036] env[61868]: DEBUG nova.compute.manager [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 555.550036] env[61868]: DEBUG nova.network.neutron [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 555.602120] env[61868]: DEBUG nova.network.neutron [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 555.690426] env[61868]: DEBUG nova.policy [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '65f6dc7b309f493f97dbc41852ca2a91', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '235d12c7819640b483f7c061fe103609', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 555.713521] env[61868]: INFO nova.compute.manager [-] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Took 1.06 seconds to deallocate network for instance. [ 555.716540] env[61868]: DEBUG nova.compute.claims [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 555.717073] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.887630] env[61868]: DEBUG oslo_concurrency.lockutils [req-a9ba8edf-b8ee-405a-9dcc-d90345fdf7bc req-76981ad8-3735-46f5-9a06-de4bf61a42bb service nova] Releasing lock "refresh_cache-38c58279-2caf-4d74-a73b-f3179e6f7f3d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 555.929588] env[61868]: DEBUG nova.network.neutron [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.047417] env[61868]: DEBUG nova.compute.manager [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 556.289259] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40e2b545-d519-43ad-8543-3465a0ab463e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.298443] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51de932e-edce-4ba6-abd5-18a72cde490e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.335590] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a586c4a2-aaae-4701-ade5-e35f6f1a1325 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.343480] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f512a648-fb89-451a-8aea-0ced1241a20b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.357647] env[61868]: DEBUG nova.compute.provider_tree [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 556.432931] env[61868]: DEBUG oslo_concurrency.lockutils [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Releasing lock "refresh_cache-c8932fff-46b9-4d1d-ab0b-cc88823bd940" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 556.434309] env[61868]: DEBUG nova.compute.manager [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 556.434309] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 556.434920] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9ce7b9da-8b94-48d3-826e-38d7d88bda28 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.446194] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb710329-f981-4abe-aa2e-95c770bdd071 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.469850] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c8932fff-46b9-4d1d-ab0b-cc88823bd940 could not be found. [ 556.469850] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 556.470560] env[61868]: INFO nova.compute.manager [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Took 0.04 seconds to destroy the instance on the hypervisor. [ 556.470560] env[61868]: DEBUG oslo.service.loopingcall [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 556.470950] env[61868]: DEBUG nova.compute.manager [-] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 556.471168] env[61868]: DEBUG nova.network.neutron [-] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 556.521341] env[61868]: DEBUG nova.network.neutron [-] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 556.865908] env[61868]: DEBUG nova.scheduler.client.report [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 557.025763] env[61868]: DEBUG nova.network.neutron [-] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.034439] env[61868]: DEBUG nova.network.neutron [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Successfully created port: 15812a32-d5eb-4fc1-8918-13a3d9e153d7 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 557.068923] env[61868]: DEBUG nova.compute.manager [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 557.108030] env[61868]: DEBUG nova.virt.hardware [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 557.108030] env[61868]: DEBUG nova.virt.hardware [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 557.108030] env[61868]: DEBUG nova.virt.hardware [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 557.108030] env[61868]: DEBUG nova.virt.hardware [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 557.108324] env[61868]: DEBUG nova.virt.hardware [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 557.108324] env[61868]: DEBUG nova.virt.hardware [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 557.108324] env[61868]: DEBUG nova.virt.hardware [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 557.108324] env[61868]: DEBUG nova.virt.hardware [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 557.108324] env[61868]: DEBUG nova.virt.hardware [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 557.110019] env[61868]: DEBUG nova.virt.hardware [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 557.110019] env[61868]: DEBUG nova.virt.hardware [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 557.110531] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96d7a999-8f11-4d0e-84ee-8dc7a9087d01 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.119426] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31fcec14-f585-4a39-9d62-c0a8225ae69c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.374146] env[61868]: DEBUG oslo_concurrency.lockutils [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.338s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 557.374146] env[61868]: DEBUG nova.compute.manager [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 557.375738] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.877s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.378706] env[61868]: INFO nova.compute.claims [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: da571582-dfe5-434b-a947-37b075b31f97] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 557.511349] env[61868]: DEBUG nova.compute.manager [req-3eb2be6b-c301-432d-89d7-545af23b7e68 req-2fe07aca-66ae-46ca-8653-46f26ff14f7e service nova] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Received event network-vif-deleted-720aed11-8a04-47a0-85cb-a0bfe9c615c1 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 557.528437] env[61868]: INFO nova.compute.manager [-] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Took 1.06 seconds to deallocate network for instance. [ 557.532805] env[61868]: DEBUG nova.compute.claims [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 557.532946] env[61868]: DEBUG oslo_concurrency.lockutils [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.796673] env[61868]: DEBUG nova.compute.manager [req-a66c6266-ffb9-4795-baba-5a428bf791a3 req-2fc22892-c153-46bc-a545-03e5b5844805 service nova] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Received event network-changed-da55d5ca-ef4c-407e-81d9-457f3207cc1e {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 557.797049] env[61868]: DEBUG nova.compute.manager [req-a66c6266-ffb9-4795-baba-5a428bf791a3 req-2fc22892-c153-46bc-a545-03e5b5844805 service nova] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Refreshing instance network info cache due to event network-changed-da55d5ca-ef4c-407e-81d9-457f3207cc1e. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 557.797141] env[61868]: DEBUG oslo_concurrency.lockutils [req-a66c6266-ffb9-4795-baba-5a428bf791a3 req-2fc22892-c153-46bc-a545-03e5b5844805 service nova] Acquiring lock "refresh_cache-c8932fff-46b9-4d1d-ab0b-cc88823bd940" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 557.797232] env[61868]: DEBUG oslo_concurrency.lockutils [req-a66c6266-ffb9-4795-baba-5a428bf791a3 req-2fc22892-c153-46bc-a545-03e5b5844805 service nova] Acquired lock "refresh_cache-c8932fff-46b9-4d1d-ab0b-cc88823bd940" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 557.797434] env[61868]: DEBUG nova.network.neutron [req-a66c6266-ffb9-4795-baba-5a428bf791a3 req-2fc22892-c153-46bc-a545-03e5b5844805 service nova] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Refreshing network info cache for port da55d5ca-ef4c-407e-81d9-457f3207cc1e {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 557.887424] env[61868]: DEBUG nova.compute.utils [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 557.892595] env[61868]: DEBUG nova.compute.manager [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Not allocating networking since 'none' was specified. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 558.074204] env[61868]: DEBUG nova.network.neutron [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Successfully created port: 9abd76d6-7797-4cac-ac77-0340c8b32da6 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 558.146430] env[61868]: DEBUG oslo_concurrency.lockutils [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] Acquiring lock "e7439426-722a-4420-8607-903cf96dff06" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.146430] env[61868]: DEBUG oslo_concurrency.lockutils [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] Lock "e7439426-722a-4420-8607-903cf96dff06" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.394487] env[61868]: DEBUG nova.compute.manager [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 558.408439] env[61868]: DEBUG nova.network.neutron [req-a66c6266-ffb9-4795-baba-5a428bf791a3 req-2fc22892-c153-46bc-a545-03e5b5844805 service nova] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 558.672859] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2e61bd3-62dc-4414-8404-21b92c8bb2b8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.681767] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae62f19d-96ad-4191-a071-218a4ee2b67d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.722655] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-168bb5ee-6f85-4f9d-ac2b-5d4dce960372 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.732713] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa025b17-1c13-43b2-825b-f948b6345fcd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.747856] env[61868]: DEBUG nova.compute.provider_tree [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 558.982542] env[61868]: DEBUG nova.network.neutron [req-a66c6266-ffb9-4795-baba-5a428bf791a3 req-2fc22892-c153-46bc-a545-03e5b5844805 service nova] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.995305] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] Acquiring lock "a93e681e-1cb1-4025-8725-4c6664d78c96" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.995605] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] Lock "a93e681e-1cb1-4025-8725-4c6664d78c96" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.003s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 559.251274] env[61868]: DEBUG nova.scheduler.client.report [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 559.281445] env[61868]: ERROR nova.compute.manager [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 286e88a8-7583-40d1-833c-9c34c435ca30, please check neutron logs for more information. [ 559.281445] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 559.281445] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 559.281445] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 559.281445] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 559.281445] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 559.281445] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 559.281445] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 559.281445] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 559.281445] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 559.281445] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 559.281445] env[61868]: ERROR nova.compute.manager raise self.value [ 559.281445] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 559.281445] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 559.281445] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 559.281445] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 559.281922] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 559.281922] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 559.281922] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 286e88a8-7583-40d1-833c-9c34c435ca30, please check neutron logs for more information. [ 559.281922] env[61868]: ERROR nova.compute.manager [ 559.281922] env[61868]: Traceback (most recent call last): [ 559.281922] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 559.281922] env[61868]: listener.cb(fileno) [ 559.281922] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 559.281922] env[61868]: result = function(*args, **kwargs) [ 559.281922] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 559.281922] env[61868]: return func(*args, **kwargs) [ 559.281922] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 559.281922] env[61868]: raise e [ 559.281922] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 559.281922] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 559.281922] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 559.281922] env[61868]: created_port_ids = self._update_ports_for_instance( [ 559.281922] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 559.281922] env[61868]: with excutils.save_and_reraise_exception(): [ 559.281922] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 559.281922] env[61868]: self.force_reraise() [ 559.281922] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 559.281922] env[61868]: raise self.value [ 559.281922] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 559.281922] env[61868]: updated_port = self._update_port( [ 559.281922] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 559.281922] env[61868]: _ensure_no_port_binding_failure(port) [ 559.281922] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 559.281922] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 559.283691] env[61868]: nova.exception.PortBindingFailed: Binding failed for port 286e88a8-7583-40d1-833c-9c34c435ca30, please check neutron logs for more information. [ 559.283691] env[61868]: Removing descriptor: 16 [ 559.283691] env[61868]: ERROR nova.compute.manager [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 286e88a8-7583-40d1-833c-9c34c435ca30, please check neutron logs for more information. [ 559.283691] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Traceback (most recent call last): [ 559.283691] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 559.283691] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] yield resources [ 559.283691] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 559.283691] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] self.driver.spawn(context, instance, image_meta, [ 559.283691] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 559.283691] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] self._vmops.spawn(context, instance, image_meta, injected_files, [ 559.283691] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 559.283691] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] vm_ref = self.build_virtual_machine(instance, [ 559.284218] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 559.284218] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] vif_infos = vmwarevif.get_vif_info(self._session, [ 559.284218] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 559.284218] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] for vif in network_info: [ 559.284218] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 559.284218] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] return self._sync_wrapper(fn, *args, **kwargs) [ 559.284218] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 559.284218] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] self.wait() [ 559.284218] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 559.284218] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] self[:] = self._gt.wait() [ 559.284218] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 559.284218] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] return self._exit_event.wait() [ 559.284218] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 559.284608] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] result = hub.switch() [ 559.284608] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 559.284608] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] return self.greenlet.switch() [ 559.284608] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 559.284608] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] result = function(*args, **kwargs) [ 559.284608] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 559.284608] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] return func(*args, **kwargs) [ 559.284608] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 559.284608] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] raise e [ 559.284608] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 559.284608] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] nwinfo = self.network_api.allocate_for_instance( [ 559.284608] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 559.284608] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] created_port_ids = self._update_ports_for_instance( [ 559.284938] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 559.284938] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] with excutils.save_and_reraise_exception(): [ 559.284938] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 559.284938] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] self.force_reraise() [ 559.284938] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 559.284938] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] raise self.value [ 559.284938] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 559.284938] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] updated_port = self._update_port( [ 559.284938] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 559.284938] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] _ensure_no_port_binding_failure(port) [ 559.284938] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 559.284938] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] raise exception.PortBindingFailed(port_id=port['id']) [ 559.285259] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] nova.exception.PortBindingFailed: Binding failed for port 286e88a8-7583-40d1-833c-9c34c435ca30, please check neutron logs for more information. [ 559.285259] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] [ 559.285259] env[61868]: INFO nova.compute.manager [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Terminating instance [ 559.288445] env[61868]: DEBUG oslo_concurrency.lockutils [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] Acquiring lock "refresh_cache-a22aee91-ad9b-4eb1-be66-315900bacf78" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 559.288600] env[61868]: DEBUG oslo_concurrency.lockutils [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] Acquired lock "refresh_cache-a22aee91-ad9b-4eb1-be66-315900bacf78" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 559.288759] env[61868]: DEBUG nova.network.neutron [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 559.357680] env[61868]: DEBUG nova.network.neutron [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Successfully created port: eef08aa3-af03-4ee4-b441-cb77756c1152 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 559.405570] env[61868]: DEBUG nova.compute.manager [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 559.442070] env[61868]: DEBUG nova.virt.hardware [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 559.442656] env[61868]: DEBUG nova.virt.hardware [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 559.442656] env[61868]: DEBUG nova.virt.hardware [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 559.442656] env[61868]: DEBUG nova.virt.hardware [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 559.442804] env[61868]: DEBUG nova.virt.hardware [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 559.442939] env[61868]: DEBUG nova.virt.hardware [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 559.444117] env[61868]: DEBUG nova.virt.hardware [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 559.444339] env[61868]: DEBUG nova.virt.hardware [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 559.444517] env[61868]: DEBUG nova.virt.hardware [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 559.444680] env[61868]: DEBUG nova.virt.hardware [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 559.444847] env[61868]: DEBUG nova.virt.hardware [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 559.446541] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-401197b6-7b44-4079-a1fd-272959e9f7c3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.456216] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80a97034-7d8d-4302-8e5a-cbeae8731a52 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.471402] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Instance VIF info [] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 559.484021] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 559.484021] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-662e0731-f222-4981-8b94-35448da7ead7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.484621] env[61868]: DEBUG oslo_concurrency.lockutils [req-a66c6266-ffb9-4795-baba-5a428bf791a3 req-2fc22892-c153-46bc-a545-03e5b5844805 service nova] Releasing lock "refresh_cache-c8932fff-46b9-4d1d-ab0b-cc88823bd940" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 559.484863] env[61868]: DEBUG nova.compute.manager [req-a66c6266-ffb9-4795-baba-5a428bf791a3 req-2fc22892-c153-46bc-a545-03e5b5844805 service nova] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Received event network-vif-deleted-da55d5ca-ef4c-407e-81d9-457f3207cc1e {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 559.493548] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Created folder: OpenStack in parent group-v4. [ 559.493729] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Creating folder: Project (e27b7d690e6c4bf99a84b03e990267ed). Parent ref: group-v281478. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 559.493996] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f0c9b6c1-7448-4eb3-97ef-8df7647303a6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.503046] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Created folder: Project (e27b7d690e6c4bf99a84b03e990267ed) in parent group-v281478. [ 559.503818] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Creating folder: Instances. Parent ref: group-v281479. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 559.503818] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0f89ce9c-94ad-41a9-bdce-51dab931b629 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.511683] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Created folder: Instances in parent group-v281479. [ 559.511924] env[61868]: DEBUG oslo.service.loopingcall [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 559.512129] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 559.512340] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0a244fe6-ae50-48b4-b8d2-00f0a136d4e8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.534037] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 559.534037] env[61868]: value = "task-1315251" [ 559.534037] env[61868]: _type = "Task" [ 559.534037] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 559.544317] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315251, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 559.756994] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.381s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 559.757678] env[61868]: DEBUG nova.compute.manager [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: da571582-dfe5-434b-a947-37b075b31f97] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 559.761221] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.912s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 559.761649] env[61868]: INFO nova.compute.claims [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] [instance: be7273d1-963e-44bf-9862-35703e37081d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 559.886239] env[61868]: DEBUG nova.network.neutron [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 559.980040] env[61868]: DEBUG nova.network.neutron [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.046091] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315251, 'name': CreateVM_Task, 'duration_secs': 0.368492} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 560.046392] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 560.047445] env[61868]: DEBUG oslo_vmware.service [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49e380c8-2eb0-4425-8ba1-fd00553bd9e4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.054235] env[61868]: DEBUG oslo_concurrency.lockutils [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 560.054439] env[61868]: DEBUG oslo_concurrency.lockutils [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 560.056290] env[61868]: DEBUG oslo_concurrency.lockutils [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 560.057608] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d9a91d69-0c4e-4437-bec3-74e6d03ae602 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.061674] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Waiting for the task: (returnval){ [ 560.061674] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]523ed923-7185-d287-fcb7-a00218972f5e" [ 560.061674] env[61868]: _type = "Task" [ 560.061674] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 560.071972] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]523ed923-7185-d287-fcb7-a00218972f5e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 560.270712] env[61868]: DEBUG nova.compute.utils [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 560.273981] env[61868]: DEBUG nova.compute.manager [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: da571582-dfe5-434b-a947-37b075b31f97] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 560.274334] env[61868]: DEBUG nova.network.neutron [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: da571582-dfe5-434b-a947-37b075b31f97] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 560.435950] env[61868]: DEBUG nova.policy [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9a68749e736f43c3a686e16b85eb964b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6f592ea4ad384a8f9669a09eb3ed9bb2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 560.484517] env[61868]: DEBUG oslo_concurrency.lockutils [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] Releasing lock "refresh_cache-a22aee91-ad9b-4eb1-be66-315900bacf78" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 560.484985] env[61868]: DEBUG nova.compute.manager [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 560.485143] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 560.485411] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cf88af63-984f-4947-9332-d2345a322a76 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.499118] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8631c7a1-a009-475f-ad67-a03bb6a0ca1f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.527912] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a22aee91-ad9b-4eb1-be66-315900bacf78 could not be found. [ 560.527912] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 560.527912] env[61868]: INFO nova.compute.manager [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Took 0.04 seconds to destroy the instance on the hypervisor. [ 560.527912] env[61868]: DEBUG oslo.service.loopingcall [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 560.529773] env[61868]: DEBUG nova.compute.manager [-] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 560.529896] env[61868]: DEBUG nova.network.neutron [-] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 560.538462] env[61868]: DEBUG oslo_concurrency.lockutils [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] Acquiring lock "448b8246-8d41-4154-bb99-b3c66d8e9475" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.538690] env[61868]: DEBUG oslo_concurrency.lockutils [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] Lock "448b8246-8d41-4154-bb99-b3c66d8e9475" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.574167] env[61868]: DEBUG oslo_concurrency.lockutils [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 560.574420] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 560.574642] env[61868]: DEBUG oslo_concurrency.lockutils [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 560.574780] env[61868]: DEBUG oslo_concurrency.lockutils [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 560.576757] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 560.579720] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5b04d4cf-1d8a-40df-88d8-9d09275535cd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.609541] env[61868]: DEBUG nova.network.neutron [-] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 560.785699] env[61868]: DEBUG nova.compute.manager [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: da571582-dfe5-434b-a947-37b075b31f97] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 561.049665] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b9be5e2-9bd4-49a0-aca5-55ed5f480473 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.066822] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-401fa07d-be44-47f9-82e1-5fa172207835 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.103693] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be3c9df1-1453-4547-9aa6-ff30f7fa40b6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.111862] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c26ec64e-b340-4390-ad83-81fd562d30a4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.116185] env[61868]: DEBUG nova.network.neutron [-] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.131032] env[61868]: DEBUG nova.compute.provider_tree [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 561.157358] env[61868]: DEBUG nova.compute.manager [req-a699917a-bd87-4829-824f-6014194c4983 req-30067a66-7602-49c1-83ad-d37f991901f3 service nova] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Received event network-changed-286e88a8-7583-40d1-833c-9c34c435ca30 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 561.157358] env[61868]: DEBUG nova.compute.manager [req-a699917a-bd87-4829-824f-6014194c4983 req-30067a66-7602-49c1-83ad-d37f991901f3 service nova] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Refreshing instance network info cache due to event network-changed-286e88a8-7583-40d1-833c-9c34c435ca30. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 561.157358] env[61868]: DEBUG oslo_concurrency.lockutils [req-a699917a-bd87-4829-824f-6014194c4983 req-30067a66-7602-49c1-83ad-d37f991901f3 service nova] Acquiring lock "refresh_cache-a22aee91-ad9b-4eb1-be66-315900bacf78" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 561.157358] env[61868]: DEBUG oslo_concurrency.lockutils [req-a699917a-bd87-4829-824f-6014194c4983 req-30067a66-7602-49c1-83ad-d37f991901f3 service nova] Acquired lock "refresh_cache-a22aee91-ad9b-4eb1-be66-315900bacf78" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 561.157358] env[61868]: DEBUG nova.network.neutron [req-a699917a-bd87-4829-824f-6014194c4983 req-30067a66-7602-49c1-83ad-d37f991901f3 service nova] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Refreshing network info cache for port 286e88a8-7583-40d1-833c-9c34c435ca30 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 561.588392] env[61868]: DEBUG nova.network.neutron [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: da571582-dfe5-434b-a947-37b075b31f97] Successfully created port: 94eb4bd3-c04e-4472-81a4-80692bb2346c {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 561.620127] env[61868]: INFO nova.compute.manager [-] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Took 1.09 seconds to deallocate network for instance. [ 561.627106] env[61868]: DEBUG nova.compute.claims [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 561.627397] env[61868]: DEBUG oslo_concurrency.lockutils [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.639173] env[61868]: DEBUG nova.scheduler.client.report [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 561.720445] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 561.720634] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 561.722018] env[61868]: DEBUG nova.network.neutron [req-a699917a-bd87-4829-824f-6014194c4983 req-30067a66-7602-49c1-83ad-d37f991901f3 service nova] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 561.728930] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cb4ad4f-a98e-492f-841e-2c82def19135 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.737398] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ad338c1-effa-44d1-b21d-36d9f96c30bf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.744132] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Waiting for the task: (returnval){ [ 561.744132] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5240fcdd-6e58-142d-7552-7cd3047901fe" [ 561.744132] env[61868]: _type = "Task" [ 561.744132] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 561.758657] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5240fcdd-6e58-142d-7552-7cd3047901fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 561.799110] env[61868]: DEBUG nova.compute.manager [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: da571582-dfe5-434b-a947-37b075b31f97] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 561.846955] env[61868]: DEBUG nova.virt.hardware [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 561.848103] env[61868]: DEBUG nova.virt.hardware [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 561.848471] env[61868]: DEBUG nova.virt.hardware [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 561.849150] env[61868]: DEBUG nova.virt.hardware [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 561.849622] env[61868]: DEBUG nova.virt.hardware [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 561.852020] env[61868]: DEBUG nova.virt.hardware [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 561.852020] env[61868]: DEBUG nova.virt.hardware [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 561.852020] env[61868]: DEBUG nova.virt.hardware [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 561.852020] env[61868]: DEBUG nova.virt.hardware [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 561.852020] env[61868]: DEBUG nova.virt.hardware [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 561.852531] env[61868]: DEBUG nova.virt.hardware [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 561.852531] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-155c426e-1be4-4945-8da4-8e3df84b9612 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.867755] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6834bcf0-1db7-4e58-a76a-a7e091d619e7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.973850] env[61868]: DEBUG oslo_concurrency.lockutils [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] Acquiring lock "2f9deec0-e0ad-4cb4-9868-9aa53cdf602d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.974134] env[61868]: DEBUG oslo_concurrency.lockutils [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] Lock "2f9deec0-e0ad-4cb4-9868-9aa53cdf602d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.001935] env[61868]: DEBUG nova.network.neutron [req-a699917a-bd87-4829-824f-6014194c4983 req-30067a66-7602-49c1-83ad-d37f991901f3 service nova] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.035442] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Acquiring lock "7651f6d2-1825-46a1-9380-732b50abd3cc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.035442] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Lock "7651f6d2-1825-46a1-9380-732b50abd3cc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.149321] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.389s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 562.150351] env[61868]: DEBUG nova.compute.manager [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] [instance: be7273d1-963e-44bf-9862-35703e37081d] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 562.158120] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 12.047s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.159155] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 562.160368] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61868) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 562.160368] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.612s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.167716] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c54a46e2-e311-4f2b-a055-521503ffd35b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.176568] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1f4e2d6-3e7b-4b92-84f6-67841535ca42 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.196806] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8db973b7-2dff-43b9-b3fe-0ccf4ae42886 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.204586] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5598654-97de-4c75-902e-c18443c51b7a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.246159] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181522MB free_disk=135GB free_vcpus=48 pci_devices=None {{(pid=61868) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 562.246740] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.258220] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Preparing fetch location {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 562.258473] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Creating directory with path [datastore1] vmware_temp/ceee87f1-4071-4028-a151-79ffe9b6d808/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 562.258708] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ca39bc76-366f-463f-982b-886531063ade {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.274806] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Created directory with path [datastore1] vmware_temp/ceee87f1-4071-4028-a151-79ffe9b6d808/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 562.274806] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Fetch image to [datastore1] vmware_temp/ceee87f1-4071-4028-a151-79ffe9b6d808/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/tmp-sparse.vmdk {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 562.274806] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Downloading image file data c4fdfa42-c9a0-41ca-801e-4e55c3d4328d to [datastore1] vmware_temp/ceee87f1-4071-4028-a151-79ffe9b6d808/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/tmp-sparse.vmdk on the data store datastore1 {{(pid=61868) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 562.277398] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe2f2c66-294f-4edc-88c7-24dd63dc5db5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.288760] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99473f85-37fa-48b1-aa15-9686bb9dedc5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.298777] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e35657b2-ef8a-44e7-a610-679ad6c7efde {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.333805] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba84f302-8a2c-4b54-a843-4a3230e4a17a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.340784] env[61868]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-c9a6367c-02b4-43ec-944c-dee850e16077 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.372960] env[61868]: DEBUG nova.virt.vmwareapi.images [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Downloading image file data c4fdfa42-c9a0-41ca-801e-4e55c3d4328d to the data store datastore1 {{(pid=61868) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 562.411073] env[61868]: DEBUG oslo_concurrency.lockutils [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Acquiring lock "8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.411186] env[61868]: DEBUG oslo_concurrency.lockutils [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Lock "8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.479628] env[61868]: DEBUG oslo_vmware.rw_handles [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/ceee87f1-4071-4028-a151-79ffe9b6d808/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61868) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 562.543743] env[61868]: DEBUG oslo_concurrency.lockutils [req-a699917a-bd87-4829-824f-6014194c4983 req-30067a66-7602-49c1-83ad-d37f991901f3 service nova] Releasing lock "refresh_cache-a22aee91-ad9b-4eb1-be66-315900bacf78" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 562.543915] env[61868]: DEBUG nova.compute.manager [req-a699917a-bd87-4829-824f-6014194c4983 req-30067a66-7602-49c1-83ad-d37f991901f3 service nova] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Received event network-vif-deleted-286e88a8-7583-40d1-833c-9c34c435ca30 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 562.669182] env[61868]: DEBUG nova.compute.utils [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 562.675560] env[61868]: DEBUG nova.compute.manager [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] [instance: be7273d1-963e-44bf-9862-35703e37081d] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 562.675744] env[61868]: DEBUG nova.network.neutron [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] [instance: be7273d1-963e-44bf-9862-35703e37081d] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 562.804665] env[61868]: ERROR nova.compute.manager [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 15812a32-d5eb-4fc1-8918-13a3d9e153d7, please check neutron logs for more information. [ 562.804665] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 562.804665] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 562.804665] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 562.804665] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 562.804665] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 562.804665] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 562.804665] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 562.804665] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 562.804665] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 562.804665] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 562.804665] env[61868]: ERROR nova.compute.manager raise self.value [ 562.804665] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 562.804665] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 562.804665] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 562.804665] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 562.806529] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 562.806529] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 562.806529] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 15812a32-d5eb-4fc1-8918-13a3d9e153d7, please check neutron logs for more information. [ 562.806529] env[61868]: ERROR nova.compute.manager [ 562.806529] env[61868]: Traceback (most recent call last): [ 562.806529] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 562.806529] env[61868]: listener.cb(fileno) [ 562.806529] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 562.806529] env[61868]: result = function(*args, **kwargs) [ 562.806529] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 562.806529] env[61868]: return func(*args, **kwargs) [ 562.806529] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 562.806529] env[61868]: raise e [ 562.806529] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 562.806529] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 562.806529] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 562.806529] env[61868]: created_port_ids = self._update_ports_for_instance( [ 562.806529] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 562.806529] env[61868]: with excutils.save_and_reraise_exception(): [ 562.806529] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 562.806529] env[61868]: self.force_reraise() [ 562.806529] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 562.806529] env[61868]: raise self.value [ 562.806529] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 562.806529] env[61868]: updated_port = self._update_port( [ 562.806529] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 562.806529] env[61868]: _ensure_no_port_binding_failure(port) [ 562.806529] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 562.806529] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 562.808858] env[61868]: nova.exception.PortBindingFailed: Binding failed for port 15812a32-d5eb-4fc1-8918-13a3d9e153d7, please check neutron logs for more information. [ 562.808858] env[61868]: Removing descriptor: 15 [ 562.808858] env[61868]: ERROR nova.compute.manager [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 15812a32-d5eb-4fc1-8918-13a3d9e153d7, please check neutron logs for more information. [ 562.808858] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Traceback (most recent call last): [ 562.808858] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 562.808858] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] yield resources [ 562.808858] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 562.808858] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] self.driver.spawn(context, instance, image_meta, [ 562.808858] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 562.808858] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 562.808858] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 562.808858] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] vm_ref = self.build_virtual_machine(instance, [ 562.809314] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 562.809314] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] vif_infos = vmwarevif.get_vif_info(self._session, [ 562.809314] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 562.809314] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] for vif in network_info: [ 562.809314] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 562.809314] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] return self._sync_wrapper(fn, *args, **kwargs) [ 562.809314] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 562.809314] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] self.wait() [ 562.809314] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 562.809314] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] self[:] = self._gt.wait() [ 562.809314] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 562.809314] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] return self._exit_event.wait() [ 562.809314] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 562.809683] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] result = hub.switch() [ 562.809683] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 562.809683] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] return self.greenlet.switch() [ 562.809683] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 562.809683] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] result = function(*args, **kwargs) [ 562.809683] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 562.809683] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] return func(*args, **kwargs) [ 562.809683] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 562.809683] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] raise e [ 562.809683] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 562.809683] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] nwinfo = self.network_api.allocate_for_instance( [ 562.809683] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 562.809683] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] created_port_ids = self._update_ports_for_instance( [ 562.810097] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 562.810097] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] with excutils.save_and_reraise_exception(): [ 562.810097] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 562.810097] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] self.force_reraise() [ 562.810097] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 562.810097] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] raise self.value [ 562.810097] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 562.810097] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] updated_port = self._update_port( [ 562.810097] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 562.810097] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] _ensure_no_port_binding_failure(port) [ 562.810097] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 562.810097] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] raise exception.PortBindingFailed(port_id=port['id']) [ 562.810455] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] nova.exception.PortBindingFailed: Binding failed for port 15812a32-d5eb-4fc1-8918-13a3d9e153d7, please check neutron logs for more information. [ 562.810455] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] [ 562.810455] env[61868]: INFO nova.compute.manager [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Terminating instance [ 562.812097] env[61868]: DEBUG oslo_concurrency.lockutils [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Acquiring lock "refresh_cache-f7fd9e57-9bf1-47d8-b77e-c99436391d1c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 562.812097] env[61868]: DEBUG oslo_concurrency.lockutils [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Acquired lock "refresh_cache-f7fd9e57-9bf1-47d8-b77e-c99436391d1c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 562.812308] env[61868]: DEBUG nova.network.neutron [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 562.996318] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbd0d2a4-298c-4d1e-b06b-88a46b08b4e4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.006791] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fad55e2e-84f4-4ef3-bd1d-23ea3432e924 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.013796] env[61868]: DEBUG nova.policy [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd6f6790551e0455e83aa2dcd864758df', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8cdcf076bdbe4a6a995b222c751dfa16', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 563.051461] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-722d71f5-c45f-4b5d-b1bb-78db881c2d3a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.061273] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c27fa2cf-ce5e-4d6c-9f18-de104c271402 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.079556] env[61868]: DEBUG nova.compute.provider_tree [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 563.145400] env[61868]: DEBUG oslo_vmware.rw_handles [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Completed reading data from the image iterator. {{(pid=61868) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 563.145603] env[61868]: DEBUG oslo_vmware.rw_handles [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Closing write handle for https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/ceee87f1-4071-4028-a151-79ffe9b6d808/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61868) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 563.176979] env[61868]: DEBUG nova.compute.manager [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] [instance: be7273d1-963e-44bf-9862-35703e37081d] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 563.289176] env[61868]: DEBUG nova.virt.vmwareapi.images [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Downloaded image file data c4fdfa42-c9a0-41ca-801e-4e55c3d4328d to vmware_temp/ceee87f1-4071-4028-a151-79ffe9b6d808/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/tmp-sparse.vmdk on the data store datastore1 {{(pid=61868) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 563.290101] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Caching image {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 563.290368] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Copying Virtual Disk [datastore1] vmware_temp/ceee87f1-4071-4028-a151-79ffe9b6d808/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/tmp-sparse.vmdk to [datastore1] vmware_temp/ceee87f1-4071-4028-a151-79ffe9b6d808/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 563.290642] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4c333947-faa9-4294-ad3e-b91d7e3977e4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.299562] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Waiting for the task: (returnval){ [ 563.299562] env[61868]: value = "task-1315252" [ 563.299562] env[61868]: _type = "Task" [ 563.299562] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 563.311655] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Task: {'id': task-1315252, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 563.417810] env[61868]: DEBUG nova.network.neutron [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 563.587114] env[61868]: DEBUG nova.scheduler.client.report [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 563.760770] env[61868]: DEBUG nova.network.neutron [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 563.813399] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Task: {'id': task-1315252, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 564.021790] env[61868]: DEBUG oslo_concurrency.lockutils [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Acquiring lock "7c2937c6-e6b7-4716-80fa-5538641bc8dc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.021790] env[61868]: DEBUG oslo_concurrency.lockutils [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Lock "7c2937c6-e6b7-4716-80fa-5538641bc8dc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.092797] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.933s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 564.093485] env[61868]: ERROR nova.compute.manager [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 47de6438-6d4d-45eb-bf20-1cdce97a1da2, please check neutron logs for more information. [ 564.093485] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Traceback (most recent call last): [ 564.093485] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 564.093485] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] self.driver.spawn(context, instance, image_meta, [ 564.093485] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 564.093485] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 564.093485] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 564.093485] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] vm_ref = self.build_virtual_machine(instance, [ 564.093485] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 564.093485] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] vif_infos = vmwarevif.get_vif_info(self._session, [ 564.093485] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 564.094131] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] for vif in network_info: [ 564.094131] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 564.094131] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] return self._sync_wrapper(fn, *args, **kwargs) [ 564.094131] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 564.094131] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] self.wait() [ 564.094131] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 564.094131] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] self[:] = self._gt.wait() [ 564.094131] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 564.094131] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] return self._exit_event.wait() [ 564.094131] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 564.094131] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] result = hub.switch() [ 564.094131] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 564.094131] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] return self.greenlet.switch() [ 564.094702] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 564.094702] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] result = function(*args, **kwargs) [ 564.094702] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 564.094702] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] return func(*args, **kwargs) [ 564.094702] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 564.094702] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] raise e [ 564.094702] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 564.094702] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] nwinfo = self.network_api.allocate_for_instance( [ 564.094702] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 564.094702] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] created_port_ids = self._update_ports_for_instance( [ 564.094702] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 564.094702] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] with excutils.save_and_reraise_exception(): [ 564.094702] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 564.095384] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] self.force_reraise() [ 564.095384] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 564.095384] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] raise self.value [ 564.095384] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 564.095384] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] updated_port = self._update_port( [ 564.095384] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 564.095384] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] _ensure_no_port_binding_failure(port) [ 564.095384] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 564.095384] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] raise exception.PortBindingFailed(port_id=port['id']) [ 564.095384] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] nova.exception.PortBindingFailed: Binding failed for port 47de6438-6d4d-45eb-bf20-1cdce97a1da2, please check neutron logs for more information. [ 564.095384] env[61868]: ERROR nova.compute.manager [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] [ 564.095722] env[61868]: DEBUG nova.compute.utils [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Binding failed for port 47de6438-6d4d-45eb-bf20-1cdce97a1da2, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 564.095722] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.372s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.098085] env[61868]: INFO nova.compute.claims [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 564.105711] env[61868]: DEBUG nova.compute.manager [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Build of instance 70adb078-edbe-4a66-b2c3-a6721bc3ffe6 was re-scheduled: Binding failed for port 47de6438-6d4d-45eb-bf20-1cdce97a1da2, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 564.105711] env[61868]: DEBUG nova.compute.manager [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 564.105711] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] Acquiring lock "refresh_cache-70adb078-edbe-4a66-b2c3-a6721bc3ffe6" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 564.105711] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] Acquired lock "refresh_cache-70adb078-edbe-4a66-b2c3-a6721bc3ffe6" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 564.107406] env[61868]: DEBUG nova.network.neutron [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 564.181990] env[61868]: DEBUG nova.network.neutron [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] [instance: be7273d1-963e-44bf-9862-35703e37081d] Successfully created port: 4340a23a-12e0-49bf-940c-0057392bef29 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 564.189240] env[61868]: DEBUG nova.compute.manager [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] [instance: be7273d1-963e-44bf-9862-35703e37081d] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 564.227366] env[61868]: DEBUG nova.virt.hardware [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 564.227600] env[61868]: DEBUG nova.virt.hardware [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 564.227747] env[61868]: DEBUG nova.virt.hardware [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 564.227921] env[61868]: DEBUG nova.virt.hardware [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 564.228240] env[61868]: DEBUG nova.virt.hardware [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 564.228632] env[61868]: DEBUG nova.virt.hardware [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 564.228873] env[61868]: DEBUG nova.virt.hardware [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 564.229074] env[61868]: DEBUG nova.virt.hardware [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 564.229247] env[61868]: DEBUG nova.virt.hardware [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 564.229406] env[61868]: DEBUG nova.virt.hardware [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 564.229573] env[61868]: DEBUG nova.virt.hardware [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 564.230897] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-197ba442-93b8-4ff7-8242-5011c2016f7e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.238063] env[61868]: DEBUG nova.compute.manager [req-f378bb68-e2c8-4eba-9395-d52fffcf0c4b req-11496166-27b2-4626-ac30-bfbccd906b2a service nova] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Received event network-changed-15812a32-d5eb-4fc1-8918-13a3d9e153d7 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 564.238063] env[61868]: DEBUG nova.compute.manager [req-f378bb68-e2c8-4eba-9395-d52fffcf0c4b req-11496166-27b2-4626-ac30-bfbccd906b2a service nova] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Refreshing instance network info cache due to event network-changed-15812a32-d5eb-4fc1-8918-13a3d9e153d7. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 564.238063] env[61868]: DEBUG oslo_concurrency.lockutils [req-f378bb68-e2c8-4eba-9395-d52fffcf0c4b req-11496166-27b2-4626-ac30-bfbccd906b2a service nova] Acquiring lock "refresh_cache-f7fd9e57-9bf1-47d8-b77e-c99436391d1c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 564.246226] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e414938-7768-4fc0-a05e-85b11c56577c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.269775] env[61868]: DEBUG oslo_concurrency.lockutils [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Releasing lock "refresh_cache-f7fd9e57-9bf1-47d8-b77e-c99436391d1c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 564.270386] env[61868]: DEBUG nova.compute.manager [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 564.270575] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 564.270853] env[61868]: DEBUG oslo_concurrency.lockutils [req-f378bb68-e2c8-4eba-9395-d52fffcf0c4b req-11496166-27b2-4626-ac30-bfbccd906b2a service nova] Acquired lock "refresh_cache-f7fd9e57-9bf1-47d8-b77e-c99436391d1c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 564.271302] env[61868]: DEBUG nova.network.neutron [req-f378bb68-e2c8-4eba-9395-d52fffcf0c4b req-11496166-27b2-4626-ac30-bfbccd906b2a service nova] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Refreshing network info cache for port 15812a32-d5eb-4fc1-8918-13a3d9e153d7 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 564.272403] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-deb7653f-3519-4c47-8e92-217ba7377cdf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.281320] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ad009ef-7566-42a5-81b6-9c5becd0dff8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.306689] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f7fd9e57-9bf1-47d8-b77e-c99436391d1c could not be found. [ 564.306896] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 564.307096] env[61868]: INFO nova.compute.manager [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 564.307325] env[61868]: DEBUG oslo.service.loopingcall [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 564.310457] env[61868]: DEBUG nova.compute.manager [-] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 564.310552] env[61868]: DEBUG nova.network.neutron [-] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 564.320095] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Task: {'id': task-1315252, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 564.490637] env[61868]: DEBUG nova.network.neutron [-] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 564.651278] env[61868]: DEBUG nova.network.neutron [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 564.825892] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Task: {'id': task-1315252, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 564.886311] env[61868]: DEBUG nova.network.neutron [req-f378bb68-e2c8-4eba-9395-d52fffcf0c4b req-11496166-27b2-4626-ac30-bfbccd906b2a service nova] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 564.925742] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] Acquiring lock "4c98c21e-2918-4b55-b2b5-db14630a36ef" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.925862] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] Lock "4c98c21e-2918-4b55-b2b5-db14630a36ef" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.963473] env[61868]: DEBUG nova.network.neutron [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 565.106796] env[61868]: DEBUG nova.network.neutron [req-f378bb68-e2c8-4eba-9395-d52fffcf0c4b req-11496166-27b2-4626-ac30-bfbccd906b2a service nova] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 565.327916] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Task: {'id': task-1315252, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 565.475853] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] Releasing lock "refresh_cache-70adb078-edbe-4a66-b2c3-a6721bc3ffe6" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 565.475853] env[61868]: DEBUG nova.compute.manager [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 565.475853] env[61868]: DEBUG nova.compute.manager [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 565.477380] env[61868]: DEBUG nova.network.neutron [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 565.502134] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cac9728c-f374-4da6-826a-57128e8d1925 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.510648] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20e684d1-0324-45a1-8341-296bc71fea84 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.514770] env[61868]: DEBUG nova.network.neutron [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 565.546309] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bed83d81-3e3d-461c-ac0d-e5122d6dd252 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.560015] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a88e5e7-b2f2-4ff0-b778-4acc6d56ab0e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.576989] env[61868]: DEBUG nova.compute.provider_tree [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 565.610126] env[61868]: DEBUG oslo_concurrency.lockutils [req-f378bb68-e2c8-4eba-9395-d52fffcf0c4b req-11496166-27b2-4626-ac30-bfbccd906b2a service nova] Releasing lock "refresh_cache-f7fd9e57-9bf1-47d8-b77e-c99436391d1c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 565.610394] env[61868]: DEBUG nova.compute.manager [req-f378bb68-e2c8-4eba-9395-d52fffcf0c4b req-11496166-27b2-4626-ac30-bfbccd906b2a service nova] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Received event network-vif-deleted-15812a32-d5eb-4fc1-8918-13a3d9e153d7 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 565.766275] env[61868]: ERROR nova.compute.manager [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 94eb4bd3-c04e-4472-81a4-80692bb2346c, please check neutron logs for more information. [ 565.766275] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 565.766275] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 565.766275] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 565.766275] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 565.766275] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 565.766275] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 565.766275] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 565.766275] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 565.766275] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 565.766275] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 565.766275] env[61868]: ERROR nova.compute.manager raise self.value [ 565.766275] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 565.766275] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 565.766275] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 565.766275] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 565.766847] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 565.766847] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 565.766847] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 94eb4bd3-c04e-4472-81a4-80692bb2346c, please check neutron logs for more information. [ 565.766847] env[61868]: ERROR nova.compute.manager [ 565.766847] env[61868]: Traceback (most recent call last): [ 565.766847] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 565.766847] env[61868]: listener.cb(fileno) [ 565.766847] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 565.766847] env[61868]: result = function(*args, **kwargs) [ 565.766847] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 565.766847] env[61868]: return func(*args, **kwargs) [ 565.766847] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 565.766847] env[61868]: raise e [ 565.766847] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 565.766847] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 565.766847] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 565.766847] env[61868]: created_port_ids = self._update_ports_for_instance( [ 565.766847] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 565.766847] env[61868]: with excutils.save_and_reraise_exception(): [ 565.766847] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 565.766847] env[61868]: self.force_reraise() [ 565.766847] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 565.766847] env[61868]: raise self.value [ 565.766847] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 565.766847] env[61868]: updated_port = self._update_port( [ 565.766847] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 565.766847] env[61868]: _ensure_no_port_binding_failure(port) [ 565.766847] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 565.766847] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 565.767662] env[61868]: nova.exception.PortBindingFailed: Binding failed for port 94eb4bd3-c04e-4472-81a4-80692bb2346c, please check neutron logs for more information. [ 565.767662] env[61868]: Removing descriptor: 16 [ 565.767662] env[61868]: ERROR nova.compute.manager [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: da571582-dfe5-434b-a947-37b075b31f97] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 94eb4bd3-c04e-4472-81a4-80692bb2346c, please check neutron logs for more information. [ 565.767662] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] Traceback (most recent call last): [ 565.767662] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 565.767662] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] yield resources [ 565.767662] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 565.767662] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] self.driver.spawn(context, instance, image_meta, [ 565.767662] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 565.767662] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] self._vmops.spawn(context, instance, image_meta, injected_files, [ 565.767662] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 565.767662] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] vm_ref = self.build_virtual_machine(instance, [ 565.768212] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 565.768212] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] vif_infos = vmwarevif.get_vif_info(self._session, [ 565.768212] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 565.768212] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] for vif in network_info: [ 565.768212] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 565.768212] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] return self._sync_wrapper(fn, *args, **kwargs) [ 565.768212] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 565.768212] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] self.wait() [ 565.768212] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 565.768212] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] self[:] = self._gt.wait() [ 565.768212] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 565.768212] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] return self._exit_event.wait() [ 565.768212] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 565.768627] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] result = hub.switch() [ 565.768627] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 565.768627] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] return self.greenlet.switch() [ 565.768627] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 565.768627] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] result = function(*args, **kwargs) [ 565.768627] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 565.768627] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] return func(*args, **kwargs) [ 565.768627] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 565.768627] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] raise e [ 565.768627] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 565.768627] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] nwinfo = self.network_api.allocate_for_instance( [ 565.768627] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 565.768627] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] created_port_ids = self._update_ports_for_instance( [ 565.768992] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 565.768992] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] with excutils.save_and_reraise_exception(): [ 565.768992] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 565.768992] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] self.force_reraise() [ 565.768992] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 565.768992] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] raise self.value [ 565.768992] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 565.768992] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] updated_port = self._update_port( [ 565.768992] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 565.768992] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] _ensure_no_port_binding_failure(port) [ 565.768992] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 565.768992] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] raise exception.PortBindingFailed(port_id=port['id']) [ 565.769326] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] nova.exception.PortBindingFailed: Binding failed for port 94eb4bd3-c04e-4472-81a4-80692bb2346c, please check neutron logs for more information. [ 565.769326] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] [ 565.769326] env[61868]: INFO nova.compute.manager [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: da571582-dfe5-434b-a947-37b075b31f97] Terminating instance [ 565.773471] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Acquiring lock "refresh_cache-da571582-dfe5-434b-a947-37b075b31f97" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 565.773471] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Acquired lock "refresh_cache-da571582-dfe5-434b-a947-37b075b31f97" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 565.773471] env[61868]: DEBUG nova.network.neutron [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: da571582-dfe5-434b-a947-37b075b31f97] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 565.823451] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Task: {'id': task-1315252, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 565.919440] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] Acquiring lock "c46a8746-b4b2-46f8-b16a-a7ceb5ff4611" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.920186] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] Lock "c46a8746-b4b2-46f8-b16a-a7ceb5ff4611" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.023247] env[61868]: DEBUG nova.network.neutron [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 566.080084] env[61868]: DEBUG nova.scheduler.client.report [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 566.300191] env[61868]: DEBUG nova.network.neutron [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: da571582-dfe5-434b-a947-37b075b31f97] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 566.323863] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Task: {'id': task-1315252, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 566.409577] env[61868]: DEBUG nova.network.neutron [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: da571582-dfe5-434b-a947-37b075b31f97] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 566.444284] env[61868]: DEBUG nova.network.neutron [-] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 566.525293] env[61868]: INFO nova.compute.manager [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] [instance: 70adb078-edbe-4a66-b2c3-a6721bc3ffe6] Took 1.05 seconds to deallocate network for instance. [ 566.584055] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.488s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 566.584574] env[61868]: DEBUG nova.compute.manager [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 566.590263] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.558s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.830656] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Task: {'id': task-1315252, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 566.839108] env[61868]: DEBUG nova.compute.manager [req-b40e607a-2c31-4968-aef2-54da6dd8fbf0 req-2c892c5a-46ba-4f67-898e-3d6aaad5c10d service nova] [instance: da571582-dfe5-434b-a947-37b075b31f97] Received event network-changed-94eb4bd3-c04e-4472-81a4-80692bb2346c {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 566.839607] env[61868]: DEBUG nova.compute.manager [req-b40e607a-2c31-4968-aef2-54da6dd8fbf0 req-2c892c5a-46ba-4f67-898e-3d6aaad5c10d service nova] [instance: da571582-dfe5-434b-a947-37b075b31f97] Refreshing instance network info cache due to event network-changed-94eb4bd3-c04e-4472-81a4-80692bb2346c. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 566.839833] env[61868]: DEBUG oslo_concurrency.lockutils [req-b40e607a-2c31-4968-aef2-54da6dd8fbf0 req-2c892c5a-46ba-4f67-898e-3d6aaad5c10d service nova] Acquiring lock "refresh_cache-da571582-dfe5-434b-a947-37b075b31f97" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 566.912781] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Releasing lock "refresh_cache-da571582-dfe5-434b-a947-37b075b31f97" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 566.915108] env[61868]: DEBUG nova.compute.manager [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: da571582-dfe5-434b-a947-37b075b31f97] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 566.915108] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: da571582-dfe5-434b-a947-37b075b31f97] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 566.915108] env[61868]: DEBUG oslo_concurrency.lockutils [req-b40e607a-2c31-4968-aef2-54da6dd8fbf0 req-2c892c5a-46ba-4f67-898e-3d6aaad5c10d service nova] Acquired lock "refresh_cache-da571582-dfe5-434b-a947-37b075b31f97" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 566.915108] env[61868]: DEBUG nova.network.neutron [req-b40e607a-2c31-4968-aef2-54da6dd8fbf0 req-2c892c5a-46ba-4f67-898e-3d6aaad5c10d service nova] [instance: da571582-dfe5-434b-a947-37b075b31f97] Refreshing network info cache for port 94eb4bd3-c04e-4472-81a4-80692bb2346c {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 566.915254] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0892fe20-a442-4ebb-9ebb-b3adbaab7412 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.924882] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ec76f51-984d-43b4-8ff7-dfeceda6e9ce {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.954880] env[61868]: INFO nova.compute.manager [-] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Took 2.64 seconds to deallocate network for instance. [ 566.955948] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: da571582-dfe5-434b-a947-37b075b31f97] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance da571582-dfe5-434b-a947-37b075b31f97 could not be found. [ 566.956119] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: da571582-dfe5-434b-a947-37b075b31f97] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 566.956450] env[61868]: INFO nova.compute.manager [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: da571582-dfe5-434b-a947-37b075b31f97] Took 0.04 seconds to destroy the instance on the hypervisor. [ 566.956450] env[61868]: DEBUG oslo.service.loopingcall [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 566.958203] env[61868]: DEBUG nova.compute.manager [-] [instance: da571582-dfe5-434b-a947-37b075b31f97] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 566.958298] env[61868]: DEBUG nova.network.neutron [-] [instance: da571582-dfe5-434b-a947-37b075b31f97] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 566.960791] env[61868]: DEBUG nova.compute.claims [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 566.960791] env[61868]: DEBUG oslo_concurrency.lockutils [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.008700] env[61868]: DEBUG nova.network.neutron [-] [instance: da571582-dfe5-434b-a947-37b075b31f97] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 567.100429] env[61868]: DEBUG nova.compute.utils [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 567.109421] env[61868]: DEBUG nova.compute.manager [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 567.109421] env[61868]: DEBUG nova.network.neutron [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 567.157667] env[61868]: ERROR nova.compute.manager [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4340a23a-12e0-49bf-940c-0057392bef29, please check neutron logs for more information. [ 567.157667] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 567.157667] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 567.157667] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 567.157667] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 567.157667] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 567.157667] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 567.157667] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 567.157667] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 567.157667] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 567.157667] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 567.157667] env[61868]: ERROR nova.compute.manager raise self.value [ 567.157667] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 567.157667] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 567.157667] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 567.157667] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 567.159227] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 567.159227] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 567.159227] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4340a23a-12e0-49bf-940c-0057392bef29, please check neutron logs for more information. [ 567.159227] env[61868]: ERROR nova.compute.manager [ 567.159227] env[61868]: Traceback (most recent call last): [ 567.159227] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 567.159227] env[61868]: listener.cb(fileno) [ 567.159227] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 567.159227] env[61868]: result = function(*args, **kwargs) [ 567.159227] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 567.159227] env[61868]: return func(*args, **kwargs) [ 567.159227] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 567.159227] env[61868]: raise e [ 567.159227] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 567.159227] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 567.159227] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 567.159227] env[61868]: created_port_ids = self._update_ports_for_instance( [ 567.159227] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 567.159227] env[61868]: with excutils.save_and_reraise_exception(): [ 567.159227] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 567.159227] env[61868]: self.force_reraise() [ 567.159227] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 567.159227] env[61868]: raise self.value [ 567.159227] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 567.159227] env[61868]: updated_port = self._update_port( [ 567.159227] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 567.159227] env[61868]: _ensure_no_port_binding_failure(port) [ 567.159227] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 567.159227] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 567.160561] env[61868]: nova.exception.PortBindingFailed: Binding failed for port 4340a23a-12e0-49bf-940c-0057392bef29, please check neutron logs for more information. [ 567.160561] env[61868]: Removing descriptor: 18 [ 567.160561] env[61868]: ERROR nova.compute.manager [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] [instance: be7273d1-963e-44bf-9862-35703e37081d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4340a23a-12e0-49bf-940c-0057392bef29, please check neutron logs for more information. [ 567.160561] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] Traceback (most recent call last): [ 567.160561] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 567.160561] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] yield resources [ 567.160561] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 567.160561] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] self.driver.spawn(context, instance, image_meta, [ 567.160561] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 567.160561] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 567.160561] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 567.160561] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] vm_ref = self.build_virtual_machine(instance, [ 567.161141] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 567.161141] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] vif_infos = vmwarevif.get_vif_info(self._session, [ 567.161141] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 567.161141] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] for vif in network_info: [ 567.161141] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 567.161141] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] return self._sync_wrapper(fn, *args, **kwargs) [ 567.161141] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 567.161141] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] self.wait() [ 567.161141] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 567.161141] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] self[:] = self._gt.wait() [ 567.161141] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 567.161141] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] return self._exit_event.wait() [ 567.161141] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 567.162253] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] result = hub.switch() [ 567.162253] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 567.162253] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] return self.greenlet.switch() [ 567.162253] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 567.162253] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] result = function(*args, **kwargs) [ 567.162253] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 567.162253] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] return func(*args, **kwargs) [ 567.162253] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 567.162253] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] raise e [ 567.162253] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 567.162253] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] nwinfo = self.network_api.allocate_for_instance( [ 567.162253] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 567.162253] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] created_port_ids = self._update_ports_for_instance( [ 567.162835] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 567.162835] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] with excutils.save_and_reraise_exception(): [ 567.162835] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 567.162835] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] self.force_reraise() [ 567.162835] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 567.162835] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] raise self.value [ 567.162835] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 567.162835] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] updated_port = self._update_port( [ 567.162835] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 567.162835] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] _ensure_no_port_binding_failure(port) [ 567.162835] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 567.162835] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] raise exception.PortBindingFailed(port_id=port['id']) [ 567.163414] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] nova.exception.PortBindingFailed: Binding failed for port 4340a23a-12e0-49bf-940c-0057392bef29, please check neutron logs for more information. [ 567.163414] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] [ 567.163414] env[61868]: INFO nova.compute.manager [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] [instance: be7273d1-963e-44bf-9862-35703e37081d] Terminating instance [ 567.163414] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] Acquiring lock "refresh_cache-be7273d1-963e-44bf-9862-35703e37081d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 567.163414] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] Acquired lock "refresh_cache-be7273d1-963e-44bf-9862-35703e37081d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 567.163414] env[61868]: DEBUG nova.network.neutron [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] [instance: be7273d1-963e-44bf-9862-35703e37081d] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 567.235471] env[61868]: DEBUG nova.policy [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '15e6b82bc6a84835811c4a8289c3eeeb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5d037080d42d42a6a590b46e8e5b1235', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 567.325767] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Task: {'id': task-1315252, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 567.448873] env[61868]: DEBUG nova.network.neutron [req-b40e607a-2c31-4968-aef2-54da6dd8fbf0 req-2c892c5a-46ba-4f67-898e-3d6aaad5c10d service nova] [instance: da571582-dfe5-434b-a947-37b075b31f97] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 567.517478] env[61868]: DEBUG nova.network.neutron [-] [instance: da571582-dfe5-434b-a947-37b075b31f97] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.568569] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ffc976b-4772-4ab2-8cfe-9fff24f2e889 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.573534] env[61868]: INFO nova.scheduler.client.report [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] Deleted allocations for instance 70adb078-edbe-4a66-b2c3-a6721bc3ffe6 [ 567.590467] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12866271-d9a0-4a1b-b65b-b17ea49ccacb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.624033] env[61868]: DEBUG nova.compute.manager [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 567.631940] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24693434-d1b4-43e7-bcda-1d39afb6123f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.640872] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52f8925b-cdca-4efe-916d-bf631f62aa36 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.655698] env[61868]: DEBUG nova.compute.provider_tree [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 567.738159] env[61868]: DEBUG nova.network.neutron [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] [instance: be7273d1-963e-44bf-9862-35703e37081d] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 567.811246] env[61868]: DEBUG nova.network.neutron [req-b40e607a-2c31-4968-aef2-54da6dd8fbf0 req-2c892c5a-46ba-4f67-898e-3d6aaad5c10d service nova] [instance: da571582-dfe5-434b-a947-37b075b31f97] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.827873] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Task: {'id': task-1315252, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 567.923227] env[61868]: DEBUG nova.network.neutron [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] [instance: be7273d1-963e-44bf-9862-35703e37081d] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.022373] env[61868]: INFO nova.compute.manager [-] [instance: da571582-dfe5-434b-a947-37b075b31f97] Took 1.06 seconds to deallocate network for instance. [ 568.025029] env[61868]: DEBUG nova.compute.claims [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: da571582-dfe5-434b-a947-37b075b31f97] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 568.025189] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.087731] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e6a83a2e-879b-4a77-895f-a3987be6b7dc tempest-FloatingIPsAssociationNegativeTestJSON-1890568799 tempest-FloatingIPsAssociationNegativeTestJSON-1890568799-project-member] Lock "70adb078-edbe-4a66-b2c3-a6721bc3ffe6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.870s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 568.160953] env[61868]: DEBUG nova.scheduler.client.report [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 568.297672] env[61868]: DEBUG nova.network.neutron [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Successfully created port: 88ef3ec5-fe82-4884-80a4-a41aab28f6a7 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 568.318746] env[61868]: DEBUG oslo_concurrency.lockutils [req-b40e607a-2c31-4968-aef2-54da6dd8fbf0 req-2c892c5a-46ba-4f67-898e-3d6aaad5c10d service nova] Releasing lock "refresh_cache-da571582-dfe5-434b-a947-37b075b31f97" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 568.319114] env[61868]: DEBUG nova.compute.manager [req-b40e607a-2c31-4968-aef2-54da6dd8fbf0 req-2c892c5a-46ba-4f67-898e-3d6aaad5c10d service nova] [instance: da571582-dfe5-434b-a947-37b075b31f97] Received event network-vif-deleted-94eb4bd3-c04e-4472-81a4-80692bb2346c {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 568.332277] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Task: {'id': task-1315252, 'name': CopyVirtualDisk_Task, 'duration_secs': 4.848353} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 568.332602] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Copied Virtual Disk [datastore1] vmware_temp/ceee87f1-4071-4028-a151-79ffe9b6d808/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/tmp-sparse.vmdk to [datastore1] vmware_temp/ceee87f1-4071-4028-a151-79ffe9b6d808/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 568.332787] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Deleting the datastore file [datastore1] vmware_temp/ceee87f1-4071-4028-a151-79ffe9b6d808/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/tmp-sparse.vmdk {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 568.333105] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c6131a6a-21be-40d5-b288-9e9e80013c7c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.340765] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Waiting for the task: (returnval){ [ 568.340765] env[61868]: value = "task-1315253" [ 568.340765] env[61868]: _type = "Task" [ 568.340765] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 568.353632] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Task: {'id': task-1315253, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 568.426813] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] Releasing lock "refresh_cache-be7273d1-963e-44bf-9862-35703e37081d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 568.427480] env[61868]: DEBUG nova.compute.manager [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] [instance: be7273d1-963e-44bf-9862-35703e37081d] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 568.427707] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] [instance: be7273d1-963e-44bf-9862-35703e37081d] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 568.428032] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-98230920-4180-4412-959b-f64354a65956 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.437067] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7892c1f-ed06-4f5f-95c6-8be64936d825 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.463277] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] [instance: be7273d1-963e-44bf-9862-35703e37081d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance be7273d1-963e-44bf-9862-35703e37081d could not be found. [ 568.463277] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] [instance: be7273d1-963e-44bf-9862-35703e37081d] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 568.463277] env[61868]: INFO nova.compute.manager [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] [instance: be7273d1-963e-44bf-9862-35703e37081d] Took 0.04 seconds to destroy the instance on the hypervisor. [ 568.463611] env[61868]: DEBUG oslo.service.loopingcall [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 568.463681] env[61868]: DEBUG nova.compute.manager [-] [instance: be7273d1-963e-44bf-9862-35703e37081d] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 568.463804] env[61868]: DEBUG nova.network.neutron [-] [instance: be7273d1-963e-44bf-9862-35703e37081d] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 568.512144] env[61868]: DEBUG nova.network.neutron [-] [instance: be7273d1-963e-44bf-9862-35703e37081d] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 568.594108] env[61868]: DEBUG nova.compute.manager [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 568.638843] env[61868]: DEBUG nova.compute.manager [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 568.669928] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.080s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 568.670599] env[61868]: ERROR nova.compute.manager [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d9a77d13-e9ba-4ee7-b899-0e84f7a76643, please check neutron logs for more information. [ 568.670599] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Traceback (most recent call last): [ 568.670599] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 568.670599] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] self.driver.spawn(context, instance, image_meta, [ 568.670599] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 568.670599] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] self._vmops.spawn(context, instance, image_meta, injected_files, [ 568.670599] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 568.670599] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] vm_ref = self.build_virtual_machine(instance, [ 568.670599] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 568.670599] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] vif_infos = vmwarevif.get_vif_info(self._session, [ 568.670599] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 568.671098] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] for vif in network_info: [ 568.671098] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 568.671098] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] return self._sync_wrapper(fn, *args, **kwargs) [ 568.671098] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 568.671098] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] self.wait() [ 568.671098] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 568.671098] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] self[:] = self._gt.wait() [ 568.671098] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 568.671098] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] return self._exit_event.wait() [ 568.671098] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 568.671098] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] result = hub.switch() [ 568.671098] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 568.671098] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] return self.greenlet.switch() [ 568.671428] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 568.671428] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] result = function(*args, **kwargs) [ 568.671428] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 568.671428] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] return func(*args, **kwargs) [ 568.671428] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 568.671428] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] raise e [ 568.671428] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 568.671428] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] nwinfo = self.network_api.allocate_for_instance( [ 568.671428] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 568.671428] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] created_port_ids = self._update_ports_for_instance( [ 568.671428] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 568.671428] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] with excutils.save_and_reraise_exception(): [ 568.671428] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.671741] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] self.force_reraise() [ 568.671741] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.671741] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] raise self.value [ 568.671741] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 568.671741] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] updated_port = self._update_port( [ 568.671741] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.671741] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] _ensure_no_port_binding_failure(port) [ 568.671741] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.671741] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] raise exception.PortBindingFailed(port_id=port['id']) [ 568.671741] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] nova.exception.PortBindingFailed: Binding failed for port d9a77d13-e9ba-4ee7-b899-0e84f7a76643, please check neutron logs for more information. [ 568.671741] env[61868]: ERROR nova.compute.manager [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] [ 568.672534] env[61868]: DEBUG nova.compute.utils [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Binding failed for port d9a77d13-e9ba-4ee7-b899-0e84f7a76643, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 568.674700] env[61868]: DEBUG nova.virt.hardware [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 568.674700] env[61868]: DEBUG nova.virt.hardware [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 568.674700] env[61868]: DEBUG nova.virt.hardware [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 568.674892] env[61868]: DEBUG nova.virt.hardware [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 568.674971] env[61868]: DEBUG nova.virt.hardware [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 568.675133] env[61868]: DEBUG nova.virt.hardware [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 568.675333] env[61868]: DEBUG nova.virt.hardware [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 568.675486] env[61868]: DEBUG nova.virt.hardware [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 568.675646] env[61868]: DEBUG nova.virt.hardware [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 568.675835] env[61868]: DEBUG nova.virt.hardware [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 568.675967] env[61868]: DEBUG nova.virt.hardware [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 568.676319] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.960s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.680524] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59cbc7e0-567e-44b7-b03f-b64afa7fa2ce {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.683527] env[61868]: DEBUG nova.compute.manager [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Build of instance 3442fbd9-83cf-4835-846c-0a32b2690598 was re-scheduled: Binding failed for port d9a77d13-e9ba-4ee7-b899-0e84f7a76643, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 568.683976] env[61868]: DEBUG nova.compute.manager [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 568.684216] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Acquiring lock "refresh_cache-3442fbd9-83cf-4835-846c-0a32b2690598" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 568.684363] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Acquired lock "refresh_cache-3442fbd9-83cf-4835-846c-0a32b2690598" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 568.684518] env[61868]: DEBUG nova.network.neutron [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 568.694718] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df43005a-bd19-4d1d-93fc-66d06c0f315b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.853511] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Task: {'id': task-1315253, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.064164} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 568.853511] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 568.853686] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Moving file from [datastore1] vmware_temp/ceee87f1-4071-4028-a151-79ffe9b6d808/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d to [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d. {{(pid=61868) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 568.853934] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-c6b982e4-96a1-4c45-8abc-57f5d311b94c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.860519] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Waiting for the task: (returnval){ [ 568.860519] env[61868]: value = "task-1315254" [ 568.860519] env[61868]: _type = "Task" [ 568.860519] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 568.869866] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Task: {'id': task-1315254, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 569.016938] env[61868]: DEBUG nova.network.neutron [-] [instance: be7273d1-963e-44bf-9862-35703e37081d] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 569.124632] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.278117] env[61868]: DEBUG nova.network.neutron [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 569.377099] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Task: {'id': task-1315254, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.031434} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 569.377099] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] File moved {{(pid=61868) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 569.377584] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Cleaning up location [datastore1] vmware_temp/ceee87f1-4071-4028-a151-79ffe9b6d808 {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 569.377933] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Deleting the datastore file [datastore1] vmware_temp/ceee87f1-4071-4028-a151-79ffe9b6d808 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 569.378308] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-03d954a7-c792-4400-a0b8-501224322cea {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.386128] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Waiting for the task: (returnval){ [ 569.386128] env[61868]: value = "task-1315255" [ 569.386128] env[61868]: _type = "Task" [ 569.386128] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 569.396679] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Task: {'id': task-1315255, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 569.519205] env[61868]: INFO nova.compute.manager [-] [instance: be7273d1-963e-44bf-9862-35703e37081d] Took 1.06 seconds to deallocate network for instance. [ 569.525022] env[61868]: DEBUG nova.compute.claims [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] [instance: be7273d1-963e-44bf-9862-35703e37081d] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 569.525022] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.547535] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18ecdd0f-97f1-4d03-b896-daaa759293d3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.555950] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba7fda4a-075f-46a6-b074-e206daf87b3a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.595270] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2875f18f-b555-45c4-ba5e-250ffdfbfd80 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.600031] env[61868]: DEBUG nova.compute.manager [req-ea26a25a-c9ef-4626-8666-42bf4560455f req-2fe68388-640a-4a81-a44f-86bdbb32419c service nova] [instance: be7273d1-963e-44bf-9862-35703e37081d] Received event network-changed-4340a23a-12e0-49bf-940c-0057392bef29 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 569.600031] env[61868]: DEBUG nova.compute.manager [req-ea26a25a-c9ef-4626-8666-42bf4560455f req-2fe68388-640a-4a81-a44f-86bdbb32419c service nova] [instance: be7273d1-963e-44bf-9862-35703e37081d] Refreshing instance network info cache due to event network-changed-4340a23a-12e0-49bf-940c-0057392bef29. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 569.600177] env[61868]: DEBUG oslo_concurrency.lockutils [req-ea26a25a-c9ef-4626-8666-42bf4560455f req-2fe68388-640a-4a81-a44f-86bdbb32419c service nova] Acquiring lock "refresh_cache-be7273d1-963e-44bf-9862-35703e37081d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 569.600340] env[61868]: DEBUG oslo_concurrency.lockutils [req-ea26a25a-c9ef-4626-8666-42bf4560455f req-2fe68388-640a-4a81-a44f-86bdbb32419c service nova] Acquired lock "refresh_cache-be7273d1-963e-44bf-9862-35703e37081d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 569.601367] env[61868]: DEBUG nova.network.neutron [req-ea26a25a-c9ef-4626-8666-42bf4560455f req-2fe68388-640a-4a81-a44f-86bdbb32419c service nova] [instance: be7273d1-963e-44bf-9862-35703e37081d] Refreshing network info cache for port 4340a23a-12e0-49bf-940c-0057392bef29 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 569.609653] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59a63c3c-3b70-4c73-9840-9e17809fbb85 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.627993] env[61868]: DEBUG nova.compute.provider_tree [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 569.685225] env[61868]: DEBUG nova.network.neutron [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 569.896938] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Task: {'id': task-1315255, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.026629} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 569.896938] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 569.897741] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-afc6ed0b-38fa-454c-bf1a-9121e67a8163 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.903497] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Waiting for the task: (returnval){ [ 569.903497] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52f96164-bc0f-d861-e75a-17dbd1e3543b" [ 569.903497] env[61868]: _type = "Task" [ 569.903497] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 569.913646] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52f96164-bc0f-d861-e75a-17dbd1e3543b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 570.131832] env[61868]: DEBUG nova.scheduler.client.report [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 570.153313] env[61868]: DEBUG nova.network.neutron [req-ea26a25a-c9ef-4626-8666-42bf4560455f req-2fe68388-640a-4a81-a44f-86bdbb32419c service nova] [instance: be7273d1-963e-44bf-9862-35703e37081d] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 570.188011] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Releasing lock "refresh_cache-3442fbd9-83cf-4835-846c-0a32b2690598" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 570.188269] env[61868]: DEBUG nova.compute.manager [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 570.188434] env[61868]: DEBUG nova.compute.manager [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 570.188601] env[61868]: DEBUG nova.network.neutron [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 570.244881] env[61868]: DEBUG nova.network.neutron [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 570.412844] env[61868]: DEBUG nova.network.neutron [req-ea26a25a-c9ef-4626-8666-42bf4560455f req-2fe68388-640a-4a81-a44f-86bdbb32419c service nova] [instance: be7273d1-963e-44bf-9862-35703e37081d] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.423403] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52f96164-bc0f-d861-e75a-17dbd1e3543b, 'name': SearchDatastore_Task, 'duration_secs': 0.00986} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 570.423723] env[61868]: DEBUG oslo_concurrency.lockutils [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 570.424099] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] ddde9ee7-5198-42e5-86b9-f51727dfbf60/ddde9ee7-5198-42e5-86b9-f51727dfbf60.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 570.425043] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-09c72c5d-4830-4123-84c4-6ff9b466a564 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.437171] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Waiting for the task: (returnval){ [ 570.437171] env[61868]: value = "task-1315256" [ 570.437171] env[61868]: _type = "Task" [ 570.437171] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 570.445483] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Task: {'id': task-1315256, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 570.640081] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.963s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.641175] env[61868]: ERROR nova.compute.manager [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 720aed11-8a04-47a0-85cb-a0bfe9c615c1, please check neutron logs for more information. [ 570.641175] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Traceback (most recent call last): [ 570.641175] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 570.641175] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] self.driver.spawn(context, instance, image_meta, [ 570.641175] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 570.641175] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 570.641175] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 570.641175] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] vm_ref = self.build_virtual_machine(instance, [ 570.641175] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 570.641175] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] vif_infos = vmwarevif.get_vif_info(self._session, [ 570.641175] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 570.641520] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] for vif in network_info: [ 570.641520] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 570.641520] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] return self._sync_wrapper(fn, *args, **kwargs) [ 570.641520] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 570.641520] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] self.wait() [ 570.641520] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 570.641520] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] self[:] = self._gt.wait() [ 570.641520] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 570.641520] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] return self._exit_event.wait() [ 570.641520] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 570.641520] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] result = hub.switch() [ 570.641520] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 570.641520] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] return self.greenlet.switch() [ 570.641916] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 570.641916] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] result = function(*args, **kwargs) [ 570.641916] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 570.641916] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] return func(*args, **kwargs) [ 570.641916] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 570.641916] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] raise e [ 570.641916] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 570.641916] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] nwinfo = self.network_api.allocate_for_instance( [ 570.641916] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 570.641916] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] created_port_ids = self._update_ports_for_instance( [ 570.641916] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 570.641916] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] with excutils.save_and_reraise_exception(): [ 570.641916] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 570.642355] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] self.force_reraise() [ 570.642355] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 570.642355] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] raise self.value [ 570.642355] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 570.642355] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] updated_port = self._update_port( [ 570.642355] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 570.642355] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] _ensure_no_port_binding_failure(port) [ 570.642355] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 570.642355] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] raise exception.PortBindingFailed(port_id=port['id']) [ 570.642355] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] nova.exception.PortBindingFailed: Binding failed for port 720aed11-8a04-47a0-85cb-a0bfe9c615c1, please check neutron logs for more information. [ 570.642355] env[61868]: ERROR nova.compute.manager [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] [ 570.642683] env[61868]: DEBUG nova.compute.utils [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Binding failed for port 720aed11-8a04-47a0-85cb-a0bfe9c615c1, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 570.645746] env[61868]: DEBUG oslo_concurrency.lockutils [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.112s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.651444] env[61868]: DEBUG nova.compute.manager [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Build of instance 38c58279-2caf-4d74-a73b-f3179e6f7f3d was re-scheduled: Binding failed for port 720aed11-8a04-47a0-85cb-a0bfe9c615c1, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 570.651936] env[61868]: DEBUG nova.compute.manager [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 570.652197] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] Acquiring lock "refresh_cache-38c58279-2caf-4d74-a73b-f3179e6f7f3d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 570.653240] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] Acquired lock "refresh_cache-38c58279-2caf-4d74-a73b-f3179e6f7f3d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 570.653240] env[61868]: DEBUG nova.network.neutron [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 570.747979] env[61868]: DEBUG nova.network.neutron [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.917171] env[61868]: DEBUG oslo_concurrency.lockutils [req-ea26a25a-c9ef-4626-8666-42bf4560455f req-2fe68388-640a-4a81-a44f-86bdbb32419c service nova] Releasing lock "refresh_cache-be7273d1-963e-44bf-9862-35703e37081d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 570.917549] env[61868]: DEBUG nova.compute.manager [req-ea26a25a-c9ef-4626-8666-42bf4560455f req-2fe68388-640a-4a81-a44f-86bdbb32419c service nova] [instance: be7273d1-963e-44bf-9862-35703e37081d] Received event network-vif-deleted-4340a23a-12e0-49bf-940c-0057392bef29 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 570.952998] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Task: {'id': task-1315256, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 571.252826] env[61868]: INFO nova.compute.manager [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: 3442fbd9-83cf-4835-846c-0a32b2690598] Took 1.06 seconds to deallocate network for instance. [ 571.259865] env[61868]: DEBUG nova.network.neutron [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 571.459037] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Task: {'id': task-1315256, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.521439} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 571.459037] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] ddde9ee7-5198-42e5-86b9-f51727dfbf60/ddde9ee7-5198-42e5-86b9-f51727dfbf60.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 571.459037] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 571.459037] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c0540031-57c3-4a46-8292-c4b7a3c5b0ba {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.465783] env[61868]: DEBUG nova.network.neutron [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 571.479584] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Waiting for the task: (returnval){ [ 571.479584] env[61868]: value = "task-1315257" [ 571.479584] env[61868]: _type = "Task" [ 571.479584] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 571.492620] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Task: {'id': task-1315257, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 571.559291] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57926dc6-8a69-400c-ae10-73f281df2118 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.571143] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5c8c529-a4f0-4b19-a8c4-4b663569f6af {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.616938] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-396417ef-744a-4b40-9d27-13933747002a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.625957] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d71dba0-4e15-4358-a72f-cfea709cb4b1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.642266] env[61868]: DEBUG nova.compute.provider_tree [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 571.970563] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] Releasing lock "refresh_cache-38c58279-2caf-4d74-a73b-f3179e6f7f3d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 571.970809] env[61868]: DEBUG nova.compute.manager [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 571.970977] env[61868]: DEBUG nova.compute.manager [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 571.971163] env[61868]: DEBUG nova.network.neutron [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 571.989310] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Task: {'id': task-1315257, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071892} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 571.990801] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 571.990801] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-680039f0-0b17-42b6-8fb0-9a75e8a2d45c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.013409] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Reconfiguring VM instance instance-00000007 to attach disk [datastore1] ddde9ee7-5198-42e5-86b9-f51727dfbf60/ddde9ee7-5198-42e5-86b9-f51727dfbf60.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 572.014617] env[61868]: DEBUG nova.network.neutron [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 572.017501] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-286e147e-6da8-43dc-9fd5-73bb35f71af3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.037985] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Waiting for the task: (returnval){ [ 572.037985] env[61868]: value = "task-1315258" [ 572.037985] env[61868]: _type = "Task" [ 572.037985] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 572.047111] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Task: {'id': task-1315258, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 572.145848] env[61868]: DEBUG nova.scheduler.client.report [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 572.313582] env[61868]: INFO nova.scheduler.client.report [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Deleted allocations for instance 3442fbd9-83cf-4835-846c-0a32b2690598 [ 572.533286] env[61868]: DEBUG nova.network.neutron [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 572.547842] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Task: {'id': task-1315258, 'name': ReconfigVM_Task, 'duration_secs': 0.285074} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 572.549516] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Reconfigured VM instance instance-00000007 to attach disk [datastore1] ddde9ee7-5198-42e5-86b9-f51727dfbf60/ddde9ee7-5198-42e5-86b9-f51727dfbf60.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 572.550569] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-adb6344d-3a7f-4e0a-9413-ee76421d6720 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.557672] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Waiting for the task: (returnval){ [ 572.557672] env[61868]: value = "task-1315259" [ 572.557672] env[61868]: _type = "Task" [ 572.557672] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 572.569819] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Task: {'id': task-1315259, 'name': Rename_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 572.652037] env[61868]: DEBUG oslo_concurrency.lockutils [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.007s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 572.653694] env[61868]: ERROR nova.compute.manager [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port da55d5ca-ef4c-407e-81d9-457f3207cc1e, please check neutron logs for more information. [ 572.653694] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Traceback (most recent call last): [ 572.653694] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 572.653694] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] self.driver.spawn(context, instance, image_meta, [ 572.653694] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 572.653694] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] self._vmops.spawn(context, instance, image_meta, injected_files, [ 572.653694] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 572.653694] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] vm_ref = self.build_virtual_machine(instance, [ 572.653694] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 572.653694] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] vif_infos = vmwarevif.get_vif_info(self._session, [ 572.653694] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 572.654026] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] for vif in network_info: [ 572.654026] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 572.654026] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] return self._sync_wrapper(fn, *args, **kwargs) [ 572.654026] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 572.654026] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] self.wait() [ 572.654026] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 572.654026] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] self[:] = self._gt.wait() [ 572.654026] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 572.654026] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] return self._exit_event.wait() [ 572.654026] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 572.654026] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] result = hub.switch() [ 572.654026] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 572.654026] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] return self.greenlet.switch() [ 572.654377] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 572.654377] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] result = function(*args, **kwargs) [ 572.654377] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 572.654377] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] return func(*args, **kwargs) [ 572.654377] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 572.654377] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] raise e [ 572.654377] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 572.654377] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] nwinfo = self.network_api.allocate_for_instance( [ 572.654377] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 572.654377] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] created_port_ids = self._update_ports_for_instance( [ 572.654377] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 572.654377] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] with excutils.save_and_reraise_exception(): [ 572.654377] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.654715] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] self.force_reraise() [ 572.654715] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.654715] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] raise self.value [ 572.654715] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 572.654715] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] updated_port = self._update_port( [ 572.654715] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.654715] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] _ensure_no_port_binding_failure(port) [ 572.654715] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.654715] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] raise exception.PortBindingFailed(port_id=port['id']) [ 572.654715] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] nova.exception.PortBindingFailed: Binding failed for port da55d5ca-ef4c-407e-81d9-457f3207cc1e, please check neutron logs for more information. [ 572.654715] env[61868]: ERROR nova.compute.manager [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] [ 572.655015] env[61868]: DEBUG nova.compute.utils [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Binding failed for port da55d5ca-ef4c-407e-81d9-457f3207cc1e, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 572.658174] env[61868]: DEBUG oslo_concurrency.lockutils [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.031s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.662511] env[61868]: DEBUG nova.compute.manager [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Build of instance c8932fff-46b9-4d1d-ab0b-cc88823bd940 was re-scheduled: Binding failed for port da55d5ca-ef4c-407e-81d9-457f3207cc1e, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 572.663287] env[61868]: DEBUG nova.compute.manager [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 572.663645] env[61868]: DEBUG oslo_concurrency.lockutils [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "refresh_cache-c8932fff-46b9-4d1d-ab0b-cc88823bd940" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 572.663920] env[61868]: DEBUG oslo_concurrency.lockutils [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquired lock "refresh_cache-c8932fff-46b9-4d1d-ab0b-cc88823bd940" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 572.664193] env[61868]: DEBUG nova.network.neutron [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 572.820457] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ec67c7df-1035-41e6-bec4-df7aea907eac tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Lock "3442fbd9-83cf-4835-846c-0a32b2690598" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.027s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 572.899454] env[61868]: ERROR nova.compute.manager [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 88ef3ec5-fe82-4884-80a4-a41aab28f6a7, please check neutron logs for more information. [ 572.899454] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 572.899454] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 572.899454] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 572.899454] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 572.899454] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 572.899454] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 572.899454] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 572.899454] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.899454] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 572.899454] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.899454] env[61868]: ERROR nova.compute.manager raise self.value [ 572.899454] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 572.899454] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 572.899454] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.899454] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 572.900206] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.900206] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 572.900206] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 88ef3ec5-fe82-4884-80a4-a41aab28f6a7, please check neutron logs for more information. [ 572.900206] env[61868]: ERROR nova.compute.manager [ 572.900206] env[61868]: Traceback (most recent call last): [ 572.900206] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 572.900206] env[61868]: listener.cb(fileno) [ 572.900206] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 572.900206] env[61868]: result = function(*args, **kwargs) [ 572.900206] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 572.900206] env[61868]: return func(*args, **kwargs) [ 572.900206] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 572.900206] env[61868]: raise e [ 572.900206] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 572.900206] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 572.900206] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 572.900206] env[61868]: created_port_ids = self._update_ports_for_instance( [ 572.900206] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 572.900206] env[61868]: with excutils.save_and_reraise_exception(): [ 572.900206] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.900206] env[61868]: self.force_reraise() [ 572.900206] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.900206] env[61868]: raise self.value [ 572.900206] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 572.900206] env[61868]: updated_port = self._update_port( [ 572.900206] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.900206] env[61868]: _ensure_no_port_binding_failure(port) [ 572.900206] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.900206] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 572.901772] env[61868]: nova.exception.PortBindingFailed: Binding failed for port 88ef3ec5-fe82-4884-80a4-a41aab28f6a7, please check neutron logs for more information. [ 572.901772] env[61868]: Removing descriptor: 15 [ 572.901772] env[61868]: ERROR nova.compute.manager [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 88ef3ec5-fe82-4884-80a4-a41aab28f6a7, please check neutron logs for more information. [ 572.901772] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Traceback (most recent call last): [ 572.901772] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 572.901772] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] yield resources [ 572.901772] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 572.901772] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] self.driver.spawn(context, instance, image_meta, [ 572.901772] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 572.901772] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] self._vmops.spawn(context, instance, image_meta, injected_files, [ 572.901772] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 572.901772] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] vm_ref = self.build_virtual_machine(instance, [ 572.902460] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 572.902460] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] vif_infos = vmwarevif.get_vif_info(self._session, [ 572.902460] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 572.902460] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] for vif in network_info: [ 572.902460] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 572.902460] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] return self._sync_wrapper(fn, *args, **kwargs) [ 572.902460] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 572.902460] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] self.wait() [ 572.902460] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 572.902460] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] self[:] = self._gt.wait() [ 572.902460] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 572.902460] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] return self._exit_event.wait() [ 572.902460] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 572.902999] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] result = hub.switch() [ 572.902999] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 572.902999] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] return self.greenlet.switch() [ 572.902999] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 572.902999] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] result = function(*args, **kwargs) [ 572.902999] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 572.902999] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] return func(*args, **kwargs) [ 572.902999] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 572.902999] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] raise e [ 572.902999] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 572.902999] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] nwinfo = self.network_api.allocate_for_instance( [ 572.902999] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 572.902999] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] created_port_ids = self._update_ports_for_instance( [ 572.904310] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 572.904310] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] with excutils.save_and_reraise_exception(): [ 572.904310] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.904310] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] self.force_reraise() [ 572.904310] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.904310] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] raise self.value [ 572.904310] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 572.904310] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] updated_port = self._update_port( [ 572.904310] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.904310] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] _ensure_no_port_binding_failure(port) [ 572.904310] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.904310] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] raise exception.PortBindingFailed(port_id=port['id']) [ 572.904780] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] nova.exception.PortBindingFailed: Binding failed for port 88ef3ec5-fe82-4884-80a4-a41aab28f6a7, please check neutron logs for more information. [ 572.904780] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] [ 572.904780] env[61868]: INFO nova.compute.manager [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Terminating instance [ 572.904780] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] Acquiring lock "refresh_cache-7965772b-4dd7-4807-8be4-67c6565f9304" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 572.906099] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] Acquired lock "refresh_cache-7965772b-4dd7-4807-8be4-67c6565f9304" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 572.906099] env[61868]: DEBUG nova.network.neutron [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 572.930500] env[61868]: DEBUG oslo_concurrency.lockutils [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] Acquiring lock "1586080c-3406-47f0-bbb9-0c50082eaf8d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.930936] env[61868]: DEBUG oslo_concurrency.lockutils [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] Lock "1586080c-3406-47f0-bbb9-0c50082eaf8d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.035987] env[61868]: INFO nova.compute.manager [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] [instance: 38c58279-2caf-4d74-a73b-f3179e6f7f3d] Took 1.06 seconds to deallocate network for instance. [ 573.077117] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Task: {'id': task-1315259, 'name': Rename_Task, 'duration_secs': 0.131565} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 573.077408] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 573.077635] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-46cb8857-505c-41ce-ae75-c20b937a6360 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.084268] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Waiting for the task: (returnval){ [ 573.084268] env[61868]: value = "task-1315260" [ 573.084268] env[61868]: _type = "Task" [ 573.084268] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 573.094403] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Task: {'id': task-1315260, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 573.196724] env[61868]: DEBUG nova.network.neutron [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 573.323288] env[61868]: DEBUG nova.compute.manager [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 573.330869] env[61868]: DEBUG nova.compute.manager [req-8670a5ff-5ae5-4fc1-a9f4-34623d36dd45 req-abbf41ed-0acc-4919-a28c-961844c6a10a service nova] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Received event network-changed-88ef3ec5-fe82-4884-80a4-a41aab28f6a7 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 573.331089] env[61868]: DEBUG nova.compute.manager [req-8670a5ff-5ae5-4fc1-a9f4-34623d36dd45 req-abbf41ed-0acc-4919-a28c-961844c6a10a service nova] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Refreshing instance network info cache due to event network-changed-88ef3ec5-fe82-4884-80a4-a41aab28f6a7. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 573.331400] env[61868]: DEBUG oslo_concurrency.lockutils [req-8670a5ff-5ae5-4fc1-a9f4-34623d36dd45 req-abbf41ed-0acc-4919-a28c-961844c6a10a service nova] Acquiring lock "refresh_cache-7965772b-4dd7-4807-8be4-67c6565f9304" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 573.504560] env[61868]: DEBUG nova.network.neutron [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 573.534685] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f0ee737-d283-471f-8225-229a609b27fa {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.545834] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57b870ec-9761-4df6-8b56-a8d43f979676 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.593576] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bbe44f7-7134-4d73-a3c5-3ae4b70abe0a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.595835] env[61868]: DEBUG nova.network.neutron [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.609382] env[61868]: DEBUG oslo_vmware.api [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Task: {'id': task-1315260, 'name': PowerOnVM_Task, 'duration_secs': 0.454507} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 573.610356] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 573.610543] env[61868]: INFO nova.compute.manager [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Took 14.20 seconds to spawn the instance on the hypervisor. [ 573.610794] env[61868]: DEBUG nova.compute.manager [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 573.614860] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-740c2fb3-f513-4ff0-824d-18fc0c558eea {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.619028] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf527d5f-f5c0-4201-863d-2864eae19ff1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.639044] env[61868]: DEBUG nova.compute.provider_tree [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 573.700681] env[61868]: DEBUG nova.network.neutron [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.869441] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.071835] env[61868]: INFO nova.scheduler.client.report [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] Deleted allocations for instance 38c58279-2caf-4d74-a73b-f3179e6f7f3d [ 574.103020] env[61868]: DEBUG oslo_concurrency.lockutils [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Releasing lock "refresh_cache-c8932fff-46b9-4d1d-ab0b-cc88823bd940" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.103020] env[61868]: DEBUG nova.compute.manager [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 574.103020] env[61868]: DEBUG nova.compute.manager [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 574.103020] env[61868]: DEBUG nova.network.neutron [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 574.122113] env[61868]: DEBUG nova.network.neutron [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 574.147454] env[61868]: DEBUG nova.scheduler.client.report [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 574.163255] env[61868]: INFO nova.compute.manager [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Took 30.17 seconds to build instance. [ 574.206496] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] Releasing lock "refresh_cache-7965772b-4dd7-4807-8be4-67c6565f9304" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.207085] env[61868]: DEBUG nova.compute.manager [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 574.207174] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 574.207481] env[61868]: DEBUG oslo_concurrency.lockutils [req-8670a5ff-5ae5-4fc1-a9f4-34623d36dd45 req-abbf41ed-0acc-4919-a28c-961844c6a10a service nova] Acquired lock "refresh_cache-7965772b-4dd7-4807-8be4-67c6565f9304" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 574.207650] env[61868]: DEBUG nova.network.neutron [req-8670a5ff-5ae5-4fc1-a9f4-34623d36dd45 req-abbf41ed-0acc-4919-a28c-961844c6a10a service nova] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Refreshing network info cache for port 88ef3ec5-fe82-4884-80a4-a41aab28f6a7 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 574.208684] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5e859f15-da85-477f-a7c7-7fbb6e3b9ffe {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.220438] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa584211-ee1f-4c4e-8b70-1de098e26a27 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.250020] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7965772b-4dd7-4807-8be4-67c6565f9304 could not be found. [ 574.250020] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 574.250020] env[61868]: INFO nova.compute.manager [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Took 0.04 seconds to destroy the instance on the hypervisor. [ 574.250020] env[61868]: DEBUG oslo.service.loopingcall [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 574.250020] env[61868]: DEBUG nova.compute.manager [-] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 574.250020] env[61868]: DEBUG nova.network.neutron [-] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 574.267203] env[61868]: DEBUG nova.network.neutron [-] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 574.581425] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9be3e5d1-ab9c-45e8-9742-9cffd9b5b982 tempest-ServerDiagnosticsNegativeTest-1495699836 tempest-ServerDiagnosticsNegativeTest-1495699836-project-member] Lock "38c58279-2caf-4d74-a73b-f3179e6f7f3d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.409s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 574.624039] env[61868]: DEBUG nova.network.neutron [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.663204] env[61868]: DEBUG oslo_concurrency.lockutils [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.003s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 574.663204] env[61868]: ERROR nova.compute.manager [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 286e88a8-7583-40d1-833c-9c34c435ca30, please check neutron logs for more information. [ 574.663204] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Traceback (most recent call last): [ 574.663204] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 574.663204] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] self.driver.spawn(context, instance, image_meta, [ 574.663204] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 574.663204] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] self._vmops.spawn(context, instance, image_meta, injected_files, [ 574.663204] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 574.663204] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] vm_ref = self.build_virtual_machine(instance, [ 574.663645] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 574.663645] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] vif_infos = vmwarevif.get_vif_info(self._session, [ 574.663645] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 574.663645] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] for vif in network_info: [ 574.663645] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 574.663645] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] return self._sync_wrapper(fn, *args, **kwargs) [ 574.663645] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 574.663645] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] self.wait() [ 574.663645] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 574.663645] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] self[:] = self._gt.wait() [ 574.663645] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 574.663645] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] return self._exit_event.wait() [ 574.663645] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 574.663979] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] result = hub.switch() [ 574.663979] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 574.663979] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] return self.greenlet.switch() [ 574.663979] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 574.663979] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] result = function(*args, **kwargs) [ 574.663979] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 574.663979] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] return func(*args, **kwargs) [ 574.663979] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 574.663979] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] raise e [ 574.663979] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 574.663979] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] nwinfo = self.network_api.allocate_for_instance( [ 574.663979] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 574.663979] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] created_port_ids = self._update_ports_for_instance( [ 574.664316] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 574.664316] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] with excutils.save_and_reraise_exception(): [ 574.664316] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 574.664316] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] self.force_reraise() [ 574.664316] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 574.664316] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] raise self.value [ 574.664316] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 574.664316] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] updated_port = self._update_port( [ 574.664316] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 574.664316] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] _ensure_no_port_binding_failure(port) [ 574.664316] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 574.664316] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] raise exception.PortBindingFailed(port_id=port['id']) [ 574.664625] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] nova.exception.PortBindingFailed: Binding failed for port 286e88a8-7583-40d1-833c-9c34c435ca30, please check neutron logs for more information. [ 574.664625] env[61868]: ERROR nova.compute.manager [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] [ 574.664625] env[61868]: DEBUG nova.compute.utils [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Binding failed for port 286e88a8-7583-40d1-833c-9c34c435ca30, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 574.664625] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 12.418s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.667987] env[61868]: DEBUG nova.compute.manager [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Build of instance a22aee91-ad9b-4eb1-be66-315900bacf78 was re-scheduled: Binding failed for port 286e88a8-7583-40d1-833c-9c34c435ca30, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 574.668223] env[61868]: DEBUG nova.compute.manager [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 574.668740] env[61868]: DEBUG oslo_concurrency.lockutils [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] Acquiring lock "refresh_cache-a22aee91-ad9b-4eb1-be66-315900bacf78" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 574.668861] env[61868]: DEBUG oslo_concurrency.lockutils [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] Acquired lock "refresh_cache-a22aee91-ad9b-4eb1-be66-315900bacf78" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 574.669934] env[61868]: DEBUG nova.network.neutron [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 574.671302] env[61868]: DEBUG oslo_concurrency.lockutils [None req-28f2b4e4-fbf1-4323-861f-cc69edd6355d tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Lock "ddde9ee7-5198-42e5-86b9-f51727dfbf60" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.687s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 574.761992] env[61868]: DEBUG nova.network.neutron [req-8670a5ff-5ae5-4fc1-a9f4-34623d36dd45 req-abbf41ed-0acc-4919-a28c-961844c6a10a service nova] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 574.770773] env[61868]: DEBUG nova.network.neutron [-] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.085051] env[61868]: DEBUG nova.compute.manager [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] [instance: e7439426-722a-4420-8607-903cf96dff06] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 575.112645] env[61868]: DEBUG nova.network.neutron [req-8670a5ff-5ae5-4fc1-a9f4-34623d36dd45 req-abbf41ed-0acc-4919-a28c-961844c6a10a service nova] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.130625] env[61868]: INFO nova.compute.manager [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: c8932fff-46b9-4d1d-ab0b-cc88823bd940] Took 1.03 seconds to deallocate network for instance. [ 575.178948] env[61868]: DEBUG nova.compute.manager [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 575.215382] env[61868]: DEBUG nova.network.neutron [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 575.280473] env[61868]: INFO nova.compute.manager [-] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Took 1.03 seconds to deallocate network for instance. [ 575.282887] env[61868]: DEBUG nova.compute.claims [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 575.283138] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.368712] env[61868]: DEBUG nova.network.neutron [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.479084] env[61868]: DEBUG nova.compute.manager [None req-3ecb753f-0fb2-4689-8dd2-693ff8301feb tempest-ServerDiagnosticsV248Test-669425227 tempest-ServerDiagnosticsV248Test-669425227-project-admin] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 575.481074] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5764b40-59d3-408a-ad24-e3603cc9b9e4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.488652] env[61868]: INFO nova.compute.manager [None req-3ecb753f-0fb2-4689-8dd2-693ff8301feb tempest-ServerDiagnosticsV248Test-669425227 tempest-ServerDiagnosticsV248Test-669425227-project-admin] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Retrieving diagnostics [ 575.492285] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbe88342-e83d-46f6-a27c-adf12cfac78d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.612912] env[61868]: DEBUG oslo_concurrency.lockutils [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.615439] env[61868]: DEBUG oslo_concurrency.lockutils [req-8670a5ff-5ae5-4fc1-a9f4-34623d36dd45 req-abbf41ed-0acc-4919-a28c-961844c6a10a service nova] Releasing lock "refresh_cache-7965772b-4dd7-4807-8be4-67c6565f9304" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 575.706329] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.875866] env[61868]: DEBUG oslo_concurrency.lockutils [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] Releasing lock "refresh_cache-a22aee91-ad9b-4eb1-be66-315900bacf78" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 575.875866] env[61868]: DEBUG nova.compute.manager [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 575.875866] env[61868]: DEBUG nova.compute.manager [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 575.875866] env[61868]: DEBUG nova.network.neutron [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 575.920550] env[61868]: DEBUG nova.network.neutron [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 576.193820] env[61868]: INFO nova.scheduler.client.report [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Deleted allocations for instance c8932fff-46b9-4d1d-ab0b-cc88823bd940 [ 576.227775] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance c8932fff-46b9-4d1d-ab0b-cc88823bd940 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 576.427406] env[61868]: DEBUG nova.network.neutron [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.566070] env[61868]: DEBUG nova.compute.manager [req-2e721a73-16ca-4719-a025-660e1fa20c9a req-d9ffe71c-0ba3-4cd9-a547-09a70e5ed1ad service nova] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Received event network-vif-deleted-88ef3ec5-fe82-4884-80a4-a41aab28f6a7 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 576.597174] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] Acquiring lock "ae5ec212-9dbe-4a11-94a3-8aa4576ca4db" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.597675] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] Lock "ae5ec212-9dbe-4a11-94a3-8aa4576ca4db" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.709761] env[61868]: DEBUG oslo_concurrency.lockutils [None req-403d632b-f531-440e-bb28-bc781d51f3b7 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "c8932fff-46b9-4d1d-ab0b-cc88823bd940" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.221s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 576.733443] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance a22aee91-ad9b-4eb1-be66-315900bacf78 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 576.733443] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance f7fd9e57-9bf1-47d8-b77e-c99436391d1c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 576.733443] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance ddde9ee7-5198-42e5-86b9-f51727dfbf60 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 576.733443] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance da571582-dfe5-434b-a947-37b075b31f97 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 576.733648] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance be7273d1-963e-44bf-9862-35703e37081d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 576.733648] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 7965772b-4dd7-4807-8be4-67c6565f9304 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 576.929894] env[61868]: INFO nova.compute.manager [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] [instance: a22aee91-ad9b-4eb1-be66-315900bacf78] Took 1.05 seconds to deallocate network for instance. [ 577.216229] env[61868]: DEBUG nova.compute.manager [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 577.241332] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance a499452b-9d48-4022-8d4c-0c66c6dde9ba has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 577.749989] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 5e812408-6f27-4b52-9564-4063f4526340 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 577.763182] env[61868]: DEBUG oslo_concurrency.lockutils [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 577.971052] env[61868]: INFO nova.scheduler.client.report [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] Deleted allocations for instance a22aee91-ad9b-4eb1-be66-315900bacf78 [ 578.139012] env[61868]: DEBUG oslo_concurrency.lockutils [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Acquiring lock "a8105842-1649-495d-b0c4-18ceda62bd57" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.139766] env[61868]: DEBUG oslo_concurrency.lockutils [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Lock "a8105842-1649-495d-b0c4-18ceda62bd57" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.253460] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance e7439426-722a-4420-8607-903cf96dff06 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 578.480846] env[61868]: DEBUG oslo_concurrency.lockutils [None req-45771a94-d8c4-4216-95fa-5abb9f66eeab tempest-ServerDiagnosticsTest-342239275 tempest-ServerDiagnosticsTest-342239275-project-member] Lock "a22aee91-ad9b-4eb1-be66-315900bacf78" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.539s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 578.758651] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance a93e681e-1cb1-4025-8725-4c6664d78c96 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 578.988300] env[61868]: DEBUG nova.compute.manager [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 579.261493] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 448b8246-8d41-4154-bb99-b3c66d8e9475 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 579.444576] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] Acquiring lock "2d1b59b1-504c-487f-93ac-642c490d6546" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.444576] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] Lock "2d1b59b1-504c-487f-93ac-642c490d6546" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.516462] env[61868]: DEBUG oslo_concurrency.lockutils [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.766062] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 580.271574] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 7651f6d2-1825-46a1-9380-732b50abd3cc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 580.299916] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Acquiring lock "3e970ecb-c842-4706-afd0-176d0088e1d3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.300517] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Lock "3e970ecb-c842-4706-afd0-176d0088e1d3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.774869] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 581.282019] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 7c2937c6-e6b7-4716-80fa-5538641bc8dc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 581.304502] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Acquiring lock "85dbaf67-049c-4a90-a3dc-67e2decc1dbf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.304739] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Lock "85dbaf67-049c-4a90-a3dc-67e2decc1dbf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 581.611075] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Acquiring lock "c979d105-12b7-485c-a136-55ffad02fe67" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.611075] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Lock "c979d105-12b7-485c-a136-55ffad02fe67" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 581.784837] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 4c98c21e-2918-4b55-b2b5-db14630a36ef has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 582.291903] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance c46a8746-b4b2-46f8-b16a-a7ceb5ff4611 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 582.317449] env[61868]: DEBUG oslo_concurrency.lockutils [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "0d80665a-c881-454a-90c7-96348e7fab74" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.317449] env[61868]: DEBUG oslo_concurrency.lockutils [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "0d80665a-c881-454a-90c7-96348e7fab74" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.005s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.798569] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 1586080c-3406-47f0-bbb9-0c50082eaf8d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 582.798569] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=61868) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 582.798569] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=61868) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 583.321687] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc2c989b-aff0-4d89-851e-b8bed8ab6126 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.330683] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef6f47f1-71ca-4c24-839d-76f7db8298cb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.367722] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91dd12e0-f2ae-4e39-b5b1-0f332059d511 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.378065] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afae4fa7-50b3-439d-8008-09e23ad80095 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.394914] env[61868]: DEBUG nova.compute.provider_tree [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 583.898505] env[61868]: DEBUG nova.scheduler.client.report [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 584.405564] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61868) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 584.405564] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 9.742s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 584.405875] env[61868]: DEBUG oslo_concurrency.lockutils [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.445s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.407478] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquiring lock "b6118327-0376-4b05-aee3-cb56d9cdf0b2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.407850] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Lock "b6118327-0376-4b05-aee3-cb56d9cdf0b2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.420531] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] Acquiring lock "c0c9e61b-ab2b-47dc-997b-570204ab4c3b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.421191] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] Lock "c0c9e61b-ab2b-47dc-997b-570204ab4c3b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.453980] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6309e5e9-966a-41f6-a2a1-8697318caba6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.469248] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-506acf94-34a5-4e0b-af84-0141f3808f38 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.474337] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquiring lock "96694bb9-fbf2-4b71-9cb5-37f81b634992" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.474337] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Lock "96694bb9-fbf2-4b71-9cb5-37f81b634992" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.506572] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9faa1246-2a49-4104-897b-e145fa21e0c6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.515352] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5e3da78-5b78-4f4c-85f0-35a75845a4ed {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.529998] env[61868]: DEBUG nova.compute.provider_tree [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 586.033954] env[61868]: DEBUG nova.scheduler.client.report [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 586.446597] env[61868]: DEBUG nova.compute.manager [None req-b26b0bbc-b358-4354-96a1-3a3d7f025efa tempest-ServerDiagnosticsV248Test-669425227 tempest-ServerDiagnosticsV248Test-669425227-project-admin] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 586.447851] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4be74d44-567e-46d9-b93d-8ac2b268915f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.456131] env[61868]: INFO nova.compute.manager [None req-b26b0bbc-b358-4354-96a1-3a3d7f025efa tempest-ServerDiagnosticsV248Test-669425227 tempest-ServerDiagnosticsV248Test-669425227-project-admin] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Retrieving diagnostics [ 586.456875] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f5dc55a-d37d-41b6-a40c-8bc4cc2ce5b4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.542024] env[61868]: DEBUG oslo_concurrency.lockutils [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.134s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 586.542024] env[61868]: ERROR nova.compute.manager [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 15812a32-d5eb-4fc1-8918-13a3d9e153d7, please check neutron logs for more information. [ 586.542024] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Traceback (most recent call last): [ 586.542024] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 586.542024] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] self.driver.spawn(context, instance, image_meta, [ 586.542024] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 586.542024] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 586.542024] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 586.542024] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] vm_ref = self.build_virtual_machine(instance, [ 586.542384] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 586.542384] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] vif_infos = vmwarevif.get_vif_info(self._session, [ 586.542384] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 586.542384] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] for vif in network_info: [ 586.542384] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 586.542384] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] return self._sync_wrapper(fn, *args, **kwargs) [ 586.542384] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 586.542384] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] self.wait() [ 586.542384] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 586.542384] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] self[:] = self._gt.wait() [ 586.542384] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 586.542384] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] return self._exit_event.wait() [ 586.542384] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 586.542807] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] result = hub.switch() [ 586.542807] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 586.542807] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] return self.greenlet.switch() [ 586.542807] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 586.542807] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] result = function(*args, **kwargs) [ 586.542807] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 586.542807] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] return func(*args, **kwargs) [ 586.542807] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 586.542807] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] raise e [ 586.542807] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 586.542807] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] nwinfo = self.network_api.allocate_for_instance( [ 586.542807] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 586.542807] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] created_port_ids = self._update_ports_for_instance( [ 586.543170] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 586.543170] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] with excutils.save_and_reraise_exception(): [ 586.543170] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 586.543170] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] self.force_reraise() [ 586.543170] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 586.543170] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] raise self.value [ 586.543170] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 586.543170] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] updated_port = self._update_port( [ 586.543170] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 586.543170] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] _ensure_no_port_binding_failure(port) [ 586.543170] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 586.543170] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] raise exception.PortBindingFailed(port_id=port['id']) [ 586.543489] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] nova.exception.PortBindingFailed: Binding failed for port 15812a32-d5eb-4fc1-8918-13a3d9e153d7, please check neutron logs for more information. [ 586.543489] env[61868]: ERROR nova.compute.manager [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] [ 586.543489] env[61868]: DEBUG nova.compute.utils [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Binding failed for port 15812a32-d5eb-4fc1-8918-13a3d9e153d7, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 586.547781] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.520s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.549329] env[61868]: DEBUG nova.compute.manager [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Build of instance f7fd9e57-9bf1-47d8-b77e-c99436391d1c was re-scheduled: Binding failed for port 15812a32-d5eb-4fc1-8918-13a3d9e153d7, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 586.551186] env[61868]: DEBUG nova.compute.manager [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 586.551186] env[61868]: DEBUG oslo_concurrency.lockutils [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Acquiring lock "refresh_cache-f7fd9e57-9bf1-47d8-b77e-c99436391d1c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 586.551186] env[61868]: DEBUG oslo_concurrency.lockutils [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Acquired lock "refresh_cache-f7fd9e57-9bf1-47d8-b77e-c99436391d1c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 586.551186] env[61868]: DEBUG nova.network.neutron [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 587.012147] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Acquiring lock "cf7b7da2-380e-4a97-b142-0dd9e5ad3c79" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.013187] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Lock "cf7b7da2-380e-4a97-b142-0dd9e5ad3c79" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.229692] env[61868]: DEBUG nova.network.neutron [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 587.378844] env[61868]: DEBUG nova.network.neutron [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.544239] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6600afe1-f843-4f01-9824-78d0f9b50aa5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.552175] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74f4b922-2929-4ca6-8cf8-1ed72f7da416 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.586056] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a89cc5f-efd5-4a85-8095-b18b2af26e8a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.597043] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8075bae-787d-4c98-8066-83d899b7b7a9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.613729] env[61868]: DEBUG nova.compute.provider_tree [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 587.886358] env[61868]: DEBUG oslo_concurrency.lockutils [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Releasing lock "refresh_cache-f7fd9e57-9bf1-47d8-b77e-c99436391d1c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 587.886358] env[61868]: DEBUG nova.compute.manager [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 587.886358] env[61868]: DEBUG nova.compute.manager [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 587.886358] env[61868]: DEBUG nova.network.neutron [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 587.922920] env[61868]: DEBUG nova.network.neutron [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 587.947385] env[61868]: DEBUG oslo_concurrency.lockutils [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Acquiring lock "ddde9ee7-5198-42e5-86b9-f51727dfbf60" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.948118] env[61868]: DEBUG oslo_concurrency.lockutils [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Lock "ddde9ee7-5198-42e5-86b9-f51727dfbf60" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.948347] env[61868]: DEBUG oslo_concurrency.lockutils [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Acquiring lock "ddde9ee7-5198-42e5-86b9-f51727dfbf60-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.948512] env[61868]: DEBUG oslo_concurrency.lockutils [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Lock "ddde9ee7-5198-42e5-86b9-f51727dfbf60-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.948667] env[61868]: DEBUG oslo_concurrency.lockutils [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Lock "ddde9ee7-5198-42e5-86b9-f51727dfbf60-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 587.951929] env[61868]: INFO nova.compute.manager [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Terminating instance [ 587.954186] env[61868]: DEBUG oslo_concurrency.lockutils [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Acquiring lock "refresh_cache-ddde9ee7-5198-42e5-86b9-f51727dfbf60" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 587.954186] env[61868]: DEBUG oslo_concurrency.lockutils [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Acquired lock "refresh_cache-ddde9ee7-5198-42e5-86b9-f51727dfbf60" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 587.954548] env[61868]: DEBUG nova.network.neutron [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 588.119089] env[61868]: DEBUG nova.scheduler.client.report [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 588.358128] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] Acquiring lock "460255fb-6589-481b-9770-0a2b548c6194" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.358406] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] Lock "460255fb-6589-481b-9770-0a2b548c6194" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.428152] env[61868]: DEBUG nova.network.neutron [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.483312] env[61868]: DEBUG nova.network.neutron [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 588.542058] env[61868]: DEBUG nova.network.neutron [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.628108] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.082s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 588.632969] env[61868]: ERROR nova.compute.manager [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: da571582-dfe5-434b-a947-37b075b31f97] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 94eb4bd3-c04e-4472-81a4-80692bb2346c, please check neutron logs for more information. [ 588.632969] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] Traceback (most recent call last): [ 588.632969] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 588.632969] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] self.driver.spawn(context, instance, image_meta, [ 588.632969] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 588.632969] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] self._vmops.spawn(context, instance, image_meta, injected_files, [ 588.632969] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 588.632969] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] vm_ref = self.build_virtual_machine(instance, [ 588.632969] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 588.632969] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] vif_infos = vmwarevif.get_vif_info(self._session, [ 588.632969] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 588.633473] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] for vif in network_info: [ 588.633473] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 588.633473] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] return self._sync_wrapper(fn, *args, **kwargs) [ 588.633473] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 588.633473] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] self.wait() [ 588.633473] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 588.633473] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] self[:] = self._gt.wait() [ 588.633473] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 588.633473] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] return self._exit_event.wait() [ 588.633473] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 588.633473] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] result = hub.switch() [ 588.633473] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 588.633473] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] return self.greenlet.switch() [ 588.633827] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 588.633827] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] result = function(*args, **kwargs) [ 588.633827] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 588.633827] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] return func(*args, **kwargs) [ 588.633827] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 588.633827] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] raise e [ 588.633827] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 588.633827] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] nwinfo = self.network_api.allocate_for_instance( [ 588.633827] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 588.633827] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] created_port_ids = self._update_ports_for_instance( [ 588.633827] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 588.633827] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] with excutils.save_and_reraise_exception(): [ 588.633827] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.634353] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] self.force_reraise() [ 588.634353] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.634353] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] raise self.value [ 588.634353] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 588.634353] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] updated_port = self._update_port( [ 588.634353] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.634353] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] _ensure_no_port_binding_failure(port) [ 588.634353] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.634353] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] raise exception.PortBindingFailed(port_id=port['id']) [ 588.634353] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] nova.exception.PortBindingFailed: Binding failed for port 94eb4bd3-c04e-4472-81a4-80692bb2346c, please check neutron logs for more information. [ 588.634353] env[61868]: ERROR nova.compute.manager [instance: da571582-dfe5-434b-a947-37b075b31f97] [ 588.634747] env[61868]: DEBUG nova.compute.utils [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: da571582-dfe5-434b-a947-37b075b31f97] Binding failed for port 94eb4bd3-c04e-4472-81a4-80692bb2346c, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 588.635398] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.511s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.637364] env[61868]: INFO nova.compute.claims [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 588.640106] env[61868]: DEBUG nova.compute.manager [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: da571582-dfe5-434b-a947-37b075b31f97] Build of instance da571582-dfe5-434b-a947-37b075b31f97 was re-scheduled: Binding failed for port 94eb4bd3-c04e-4472-81a4-80692bb2346c, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 588.640594] env[61868]: DEBUG nova.compute.manager [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: da571582-dfe5-434b-a947-37b075b31f97] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 588.641478] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Acquiring lock "refresh_cache-da571582-dfe5-434b-a947-37b075b31f97" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 588.641478] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Acquired lock "refresh_cache-da571582-dfe5-434b-a947-37b075b31f97" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.641478] env[61868]: DEBUG nova.network.neutron [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: da571582-dfe5-434b-a947-37b075b31f97] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 588.932842] env[61868]: INFO nova.compute.manager [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: f7fd9e57-9bf1-47d8-b77e-c99436391d1c] Took 1.05 seconds to deallocate network for instance. [ 589.045538] env[61868]: DEBUG oslo_concurrency.lockutils [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Releasing lock "refresh_cache-ddde9ee7-5198-42e5-86b9-f51727dfbf60" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 589.045538] env[61868]: DEBUG nova.compute.manager [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 589.045538] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 589.047309] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc06c7e9-1bfe-4400-a340-f638c7ed47e6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.059214] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 589.063020] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c08d27ab-f8a3-4459-923d-4248c53a22c4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.066836] env[61868]: DEBUG oslo_vmware.api [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Waiting for the task: (returnval){ [ 589.066836] env[61868]: value = "task-1315261" [ 589.066836] env[61868]: _type = "Task" [ 589.066836] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 589.075838] env[61868]: DEBUG oslo_vmware.api [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Task: {'id': task-1315261, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 589.172277] env[61868]: DEBUG nova.network.neutron [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: da571582-dfe5-434b-a947-37b075b31f97] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 589.241821] env[61868]: DEBUG nova.network.neutron [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: da571582-dfe5-434b-a947-37b075b31f97] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.579823] env[61868]: DEBUG oslo_vmware.api [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Task: {'id': task-1315261, 'name': PowerOffVM_Task, 'duration_secs': 0.124828} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 589.580589] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 589.580780] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 589.581435] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-28b4912f-02e0-40c2-b033-6c6d4f6d2d22 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.609345] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 589.609345] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 589.609345] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Deleting the datastore file [datastore1] ddde9ee7-5198-42e5-86b9-f51727dfbf60 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 589.609562] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a4cf8c1a-6cbf-4a5f-8176-c19a5c2c6452 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.619127] env[61868]: DEBUG oslo_vmware.api [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Waiting for the task: (returnval){ [ 589.619127] env[61868]: value = "task-1315263" [ 589.619127] env[61868]: _type = "Task" [ 589.619127] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 589.628026] env[61868]: DEBUG oslo_vmware.api [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Task: {'id': task-1315263, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 589.743693] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Releasing lock "refresh_cache-da571582-dfe5-434b-a947-37b075b31f97" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 589.743693] env[61868]: DEBUG nova.compute.manager [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 589.744124] env[61868]: DEBUG nova.compute.manager [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: da571582-dfe5-434b-a947-37b075b31f97] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 589.744124] env[61868]: DEBUG nova.network.neutron [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: da571582-dfe5-434b-a947-37b075b31f97] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 589.778618] env[61868]: DEBUG nova.network.neutron [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: da571582-dfe5-434b-a947-37b075b31f97] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 589.983730] env[61868]: INFO nova.scheduler.client.report [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Deleted allocations for instance f7fd9e57-9bf1-47d8-b77e-c99436391d1c [ 590.135971] env[61868]: DEBUG oslo_vmware.api [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Task: {'id': task-1315263, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.331566} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 590.138932] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 590.139235] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 590.139668] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 590.139740] env[61868]: INFO nova.compute.manager [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Took 1.09 seconds to destroy the instance on the hypervisor. [ 590.139961] env[61868]: DEBUG oslo.service.loopingcall [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 590.140455] env[61868]: DEBUG nova.compute.manager [-] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 590.140554] env[61868]: DEBUG nova.network.neutron [-] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 590.188764] env[61868]: DEBUG nova.network.neutron [-] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 590.196873] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4265c405-ca48-4ed8-9917-c8f4ee54c9cc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.206846] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5c092eb-f54f-4458-bc33-ff0b1cfe2332 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.248134] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2e5e444-f4db-4aaf-9008-581b72fbd533 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.256538] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-612d6e66-3802-4837-a519-8fabf2281be8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.274237] env[61868]: DEBUG nova.compute.provider_tree [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 590.283585] env[61868]: DEBUG nova.network.neutron [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: da571582-dfe5-434b-a947-37b075b31f97] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.503361] env[61868]: DEBUG oslo_concurrency.lockutils [None req-aea9f6a3-38c7-4c18-a9dc-4cb0a4085f56 tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Lock "f7fd9e57-9bf1-47d8-b77e-c99436391d1c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.661s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 590.691919] env[61868]: DEBUG nova.network.neutron [-] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.777704] env[61868]: DEBUG nova.scheduler.client.report [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 590.787282] env[61868]: INFO nova.compute.manager [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: da571582-dfe5-434b-a947-37b075b31f97] Took 1.04 seconds to deallocate network for instance. [ 591.013424] env[61868]: DEBUG nova.compute.manager [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 591.196382] env[61868]: INFO nova.compute.manager [-] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Took 1.06 seconds to deallocate network for instance. [ 591.287866] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.651s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 591.287866] env[61868]: DEBUG nova.compute.manager [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 591.291220] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.769s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.546573] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.704529] env[61868]: DEBUG oslo_concurrency.lockutils [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.798975] env[61868]: DEBUG nova.compute.utils [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 591.801415] env[61868]: DEBUG nova.compute.manager [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 591.801415] env[61868]: DEBUG nova.network.neutron [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 591.839258] env[61868]: INFO nova.scheduler.client.report [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Deleted allocations for instance da571582-dfe5-434b-a947-37b075b31f97 [ 591.955858] env[61868]: DEBUG nova.policy [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7f3806034bf04c98acd8613f5ea5a710', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '200f3f2b1a054e4eb1f9e05a4ba5b038', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 592.310751] env[61868]: DEBUG nova.compute.manager [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 592.349949] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2517bb83-1588-4a92-8b23-b9e5f7b1ba5a tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Lock "da571582-dfe5-434b-a947-37b075b31f97" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.883s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 592.373401] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60ff1b4e-2bc6-4101-95a7-d92aa9a0eac5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.389578] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f844b7ab-a56b-4194-990f-7cd56cb0b96a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.423692] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53f0e195-1333-4340-addf-287f2d5d475e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.432793] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff41e5ea-a1cc-4e3b-bc2d-96cc6e93ec45 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.451536] env[61868]: DEBUG nova.compute.provider_tree [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 592.700195] env[61868]: DEBUG nova.network.neutron [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Successfully created port: d78df532-d2e6-447c-b5d8-9eecf145522c {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 592.856645] env[61868]: DEBUG nova.compute.manager [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 592.959143] env[61868]: DEBUG nova.scheduler.client.report [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 593.325048] env[61868]: DEBUG nova.compute.manager [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 593.357394] env[61868]: DEBUG nova.virt.hardware [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 593.357633] env[61868]: DEBUG nova.virt.hardware [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 593.357781] env[61868]: DEBUG nova.virt.hardware [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 593.357950] env[61868]: DEBUG nova.virt.hardware [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 593.358101] env[61868]: DEBUG nova.virt.hardware [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 593.358242] env[61868]: DEBUG nova.virt.hardware [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 593.358451] env[61868]: DEBUG nova.virt.hardware [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 593.358601] env[61868]: DEBUG nova.virt.hardware [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 593.358753] env[61868]: DEBUG nova.virt.hardware [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 593.358905] env[61868]: DEBUG nova.virt.hardware [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 593.359130] env[61868]: DEBUG nova.virt.hardware [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 593.359984] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43c0db27-bede-4f55-87af-56243a76fd4b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.371319] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea7b2712-7347-48e7-99bb-860aa8e9067b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.389934] env[61868]: DEBUG oslo_concurrency.lockutils [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.465082] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.174s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 593.469018] env[61868]: ERROR nova.compute.manager [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] [instance: be7273d1-963e-44bf-9862-35703e37081d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4340a23a-12e0-49bf-940c-0057392bef29, please check neutron logs for more information. [ 593.469018] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] Traceback (most recent call last): [ 593.469018] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 593.469018] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] self.driver.spawn(context, instance, image_meta, [ 593.469018] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 593.469018] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 593.469018] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 593.469018] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] vm_ref = self.build_virtual_machine(instance, [ 593.469018] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 593.469018] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] vif_infos = vmwarevif.get_vif_info(self._session, [ 593.469018] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 593.469439] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] for vif in network_info: [ 593.469439] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 593.469439] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] return self._sync_wrapper(fn, *args, **kwargs) [ 593.469439] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 593.469439] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] self.wait() [ 593.469439] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 593.469439] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] self[:] = self._gt.wait() [ 593.469439] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 593.469439] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] return self._exit_event.wait() [ 593.469439] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 593.469439] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] result = hub.switch() [ 593.469439] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 593.469439] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] return self.greenlet.switch() [ 593.469910] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 593.469910] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] result = function(*args, **kwargs) [ 593.469910] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 593.469910] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] return func(*args, **kwargs) [ 593.469910] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 593.469910] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] raise e [ 593.469910] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 593.469910] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] nwinfo = self.network_api.allocate_for_instance( [ 593.469910] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 593.469910] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] created_port_ids = self._update_ports_for_instance( [ 593.469910] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 593.469910] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] with excutils.save_and_reraise_exception(): [ 593.469910] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 593.470311] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] self.force_reraise() [ 593.470311] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 593.470311] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] raise self.value [ 593.470311] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 593.470311] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] updated_port = self._update_port( [ 593.470311] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 593.470311] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] _ensure_no_port_binding_failure(port) [ 593.470311] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 593.470311] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] raise exception.PortBindingFailed(port_id=port['id']) [ 593.470311] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] nova.exception.PortBindingFailed: Binding failed for port 4340a23a-12e0-49bf-940c-0057392bef29, please check neutron logs for more information. [ 593.470311] env[61868]: ERROR nova.compute.manager [instance: be7273d1-963e-44bf-9862-35703e37081d] [ 593.470642] env[61868]: DEBUG nova.compute.utils [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] [instance: be7273d1-963e-44bf-9862-35703e37081d] Binding failed for port 4340a23a-12e0-49bf-940c-0057392bef29, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 593.470642] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.601s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.474024] env[61868]: INFO nova.compute.claims [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 593.476107] env[61868]: DEBUG nova.compute.manager [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] [instance: be7273d1-963e-44bf-9862-35703e37081d] Build of instance be7273d1-963e-44bf-9862-35703e37081d was re-scheduled: Binding failed for port 4340a23a-12e0-49bf-940c-0057392bef29, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 593.476300] env[61868]: DEBUG nova.compute.manager [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] [instance: be7273d1-963e-44bf-9862-35703e37081d] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 593.476528] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] Acquiring lock "refresh_cache-be7273d1-963e-44bf-9862-35703e37081d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 593.479739] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] Acquired lock "refresh_cache-be7273d1-963e-44bf-9862-35703e37081d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.479739] env[61868]: DEBUG nova.network.neutron [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] [instance: be7273d1-963e-44bf-9862-35703e37081d] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 594.027260] env[61868]: DEBUG nova.network.neutron [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] [instance: be7273d1-963e-44bf-9862-35703e37081d] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 594.125885] env[61868]: DEBUG nova.network.neutron [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] [instance: be7273d1-963e-44bf-9862-35703e37081d] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.628351] env[61868]: ERROR nova.compute.manager [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d78df532-d2e6-447c-b5d8-9eecf145522c, please check neutron logs for more information. [ 594.628351] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 594.628351] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 594.628351] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 594.628351] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 594.628351] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 594.628351] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 594.628351] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 594.628351] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.628351] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 594.628351] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.628351] env[61868]: ERROR nova.compute.manager raise self.value [ 594.628351] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 594.628351] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 594.628351] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.628351] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 594.629183] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.629183] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 594.629183] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d78df532-d2e6-447c-b5d8-9eecf145522c, please check neutron logs for more information. [ 594.629183] env[61868]: ERROR nova.compute.manager [ 594.629183] env[61868]: Traceback (most recent call last): [ 594.629183] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 594.629183] env[61868]: listener.cb(fileno) [ 594.629183] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 594.629183] env[61868]: result = function(*args, **kwargs) [ 594.629183] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 594.629183] env[61868]: return func(*args, **kwargs) [ 594.629183] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 594.629183] env[61868]: raise e [ 594.629183] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 594.629183] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 594.629183] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 594.629183] env[61868]: created_port_ids = self._update_ports_for_instance( [ 594.629183] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 594.629183] env[61868]: with excutils.save_and_reraise_exception(): [ 594.629183] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.629183] env[61868]: self.force_reraise() [ 594.629183] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.629183] env[61868]: raise self.value [ 594.629183] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 594.629183] env[61868]: updated_port = self._update_port( [ 594.629183] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.629183] env[61868]: _ensure_no_port_binding_failure(port) [ 594.629183] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.629183] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 594.630093] env[61868]: nova.exception.PortBindingFailed: Binding failed for port d78df532-d2e6-447c-b5d8-9eecf145522c, please check neutron logs for more information. [ 594.630093] env[61868]: Removing descriptor: 16 [ 594.630093] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] Releasing lock "refresh_cache-be7273d1-963e-44bf-9862-35703e37081d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 594.630093] env[61868]: DEBUG nova.compute.manager [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 594.630093] env[61868]: DEBUG nova.compute.manager [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] [instance: be7273d1-963e-44bf-9862-35703e37081d] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 594.630093] env[61868]: DEBUG nova.network.neutron [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] [instance: be7273d1-963e-44bf-9862-35703e37081d] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 594.632883] env[61868]: ERROR nova.compute.manager [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d78df532-d2e6-447c-b5d8-9eecf145522c, please check neutron logs for more information. [ 594.632883] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Traceback (most recent call last): [ 594.632883] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 594.632883] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] yield resources [ 594.632883] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 594.632883] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] self.driver.spawn(context, instance, image_meta, [ 594.632883] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 594.632883] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] self._vmops.spawn(context, instance, image_meta, injected_files, [ 594.632883] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 594.632883] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] vm_ref = self.build_virtual_machine(instance, [ 594.632883] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 594.633515] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] vif_infos = vmwarevif.get_vif_info(self._session, [ 594.633515] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 594.633515] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] for vif in network_info: [ 594.633515] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 594.633515] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] return self._sync_wrapper(fn, *args, **kwargs) [ 594.633515] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 594.633515] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] self.wait() [ 594.633515] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 594.633515] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] self[:] = self._gt.wait() [ 594.633515] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 594.633515] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] return self._exit_event.wait() [ 594.633515] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 594.633515] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] result = hub.switch() [ 594.634024] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 594.634024] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] return self.greenlet.switch() [ 594.634024] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 594.634024] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] result = function(*args, **kwargs) [ 594.634024] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 594.634024] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] return func(*args, **kwargs) [ 594.634024] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 594.634024] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] raise e [ 594.634024] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 594.634024] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] nwinfo = self.network_api.allocate_for_instance( [ 594.634024] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 594.634024] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] created_port_ids = self._update_ports_for_instance( [ 594.634024] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 594.634399] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] with excutils.save_and_reraise_exception(): [ 594.634399] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.634399] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] self.force_reraise() [ 594.634399] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.634399] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] raise self.value [ 594.634399] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 594.634399] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] updated_port = self._update_port( [ 594.634399] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.634399] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] _ensure_no_port_binding_failure(port) [ 594.634399] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.634399] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] raise exception.PortBindingFailed(port_id=port['id']) [ 594.634399] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] nova.exception.PortBindingFailed: Binding failed for port d78df532-d2e6-447c-b5d8-9eecf145522c, please check neutron logs for more information. [ 594.634399] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] [ 594.634945] env[61868]: INFO nova.compute.manager [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Terminating instance [ 594.634945] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Acquiring lock "refresh_cache-a499452b-9d48-4022-8d4c-0c66c6dde9ba" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 594.634945] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Acquired lock "refresh_cache-a499452b-9d48-4022-8d4c-0c66c6dde9ba" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 594.634945] env[61868]: DEBUG nova.network.neutron [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 594.661046] env[61868]: DEBUG nova.network.neutron [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] [instance: be7273d1-963e-44bf-9862-35703e37081d] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 594.777471] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.777582] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.786973] env[61868]: DEBUG nova.compute.manager [req-45d66967-30a8-4e01-9180-5903e34e3694 req-a21efd7c-efc9-4f8e-8b61-3190d532530b service nova] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Received event network-changed-d78df532-d2e6-447c-b5d8-9eecf145522c {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 594.787238] env[61868]: DEBUG nova.compute.manager [req-45d66967-30a8-4e01-9180-5903e34e3694 req-a21efd7c-efc9-4f8e-8b61-3190d532530b service nova] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Refreshing instance network info cache due to event network-changed-d78df532-d2e6-447c-b5d8-9eecf145522c. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 594.787366] env[61868]: DEBUG oslo_concurrency.lockutils [req-45d66967-30a8-4e01-9180-5903e34e3694 req-a21efd7c-efc9-4f8e-8b61-3190d532530b service nova] Acquiring lock "refresh_cache-a499452b-9d48-4022-8d4c-0c66c6dde9ba" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 595.021496] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adca2b19-1691-47aa-bfd1-26d4c2f79210 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.029868] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c754a3e9-5545-43a9-acab-d28ad8171698 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.064595] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e566efb-065f-44b1-a522-715161cd7664 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.072276] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ded5920-4343-4bbf-b6be-0f739e8dc634 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.092142] env[61868]: DEBUG nova.compute.provider_tree [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 595.161725] env[61868]: DEBUG nova.network.neutron [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 595.169674] env[61868]: DEBUG nova.network.neutron [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] [instance: be7273d1-963e-44bf-9862-35703e37081d] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.273461] env[61868]: DEBUG nova.network.neutron [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.595106] env[61868]: DEBUG nova.scheduler.client.report [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 595.672508] env[61868]: INFO nova.compute.manager [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] [instance: be7273d1-963e-44bf-9862-35703e37081d] Took 1.04 seconds to deallocate network for instance. [ 595.780208] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Releasing lock "refresh_cache-a499452b-9d48-4022-8d4c-0c66c6dde9ba" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 595.780556] env[61868]: DEBUG nova.compute.manager [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 595.780750] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 595.781157] env[61868]: DEBUG oslo_concurrency.lockutils [req-45d66967-30a8-4e01-9180-5903e34e3694 req-a21efd7c-efc9-4f8e-8b61-3190d532530b service nova] Acquired lock "refresh_cache-a499452b-9d48-4022-8d4c-0c66c6dde9ba" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.781233] env[61868]: DEBUG nova.network.neutron [req-45d66967-30a8-4e01-9180-5903e34e3694 req-a21efd7c-efc9-4f8e-8b61-3190d532530b service nova] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Refreshing network info cache for port d78df532-d2e6-447c-b5d8-9eecf145522c {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 595.782310] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c4cbbd06-420b-4fa5-b06d-da8ed6044ecb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.793810] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31cf01cc-9b0e-4b2c-b145-7866c1318aa1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.823180] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a499452b-9d48-4022-8d4c-0c66c6dde9ba could not be found. [ 595.823873] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 595.824053] env[61868]: INFO nova.compute.manager [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Took 0.04 seconds to destroy the instance on the hypervisor. [ 595.825843] env[61868]: DEBUG oslo.service.loopingcall [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 595.825843] env[61868]: DEBUG nova.compute.manager [-] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 595.825843] env[61868]: DEBUG nova.network.neutron [-] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 595.845980] env[61868]: DEBUG nova.network.neutron [-] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 596.101221] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.630s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 596.101648] env[61868]: DEBUG nova.compute.manager [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 596.105307] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.822s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.310293] env[61868]: DEBUG nova.network.neutron [req-45d66967-30a8-4e01-9180-5903e34e3694 req-a21efd7c-efc9-4f8e-8b61-3190d532530b service nova] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 596.350455] env[61868]: DEBUG nova.network.neutron [-] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.393649] env[61868]: DEBUG nova.network.neutron [req-45d66967-30a8-4e01-9180-5903e34e3694 req-a21efd7c-efc9-4f8e-8b61-3190d532530b service nova] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.538943] env[61868]: DEBUG oslo_concurrency.lockutils [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Acquiring lock "b4a7ff58-0f48-4942-a270-de6910ff0c0c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.539187] env[61868]: DEBUG oslo_concurrency.lockutils [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Lock "b4a7ff58-0f48-4942-a270-de6910ff0c0c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.615729] env[61868]: DEBUG nova.compute.utils [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 596.617789] env[61868]: DEBUG nova.compute.manager [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 596.617789] env[61868]: DEBUG nova.network.neutron [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] [instance: 5e812408-6f27-4b52-9564-4063f4526340] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 596.674987] env[61868]: DEBUG nova.policy [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e53ed8f142c54ff5a86804c2bb0cea6d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2c7db2f84d1e4ccd9ae56a96fc392fb1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 596.702145] env[61868]: INFO nova.scheduler.client.report [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] Deleted allocations for instance be7273d1-963e-44bf-9862-35703e37081d [ 596.856911] env[61868]: INFO nova.compute.manager [-] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Took 1.03 seconds to deallocate network for instance. [ 596.857954] env[61868]: DEBUG nova.compute.claims [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 596.858303] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.896050] env[61868]: DEBUG oslo_concurrency.lockutils [req-45d66967-30a8-4e01-9180-5903e34e3694 req-a21efd7c-efc9-4f8e-8b61-3190d532530b service nova] Releasing lock "refresh_cache-a499452b-9d48-4022-8d4c-0c66c6dde9ba" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 596.912974] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Acquiring lock "92d3557c-655f-45d9-ae7d-bbe62567e1e2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.913246] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Lock "92d3557c-655f-45d9-ae7d-bbe62567e1e2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.029754] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3663162e-5733-40b8-924b-402320bb50a2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.038346] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7297c719-4928-46b7-b3b5-4ba7c2bb217b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.070242] env[61868]: DEBUG nova.network.neutron [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Successfully created port: e00fd5a7-bf99-4fb6-a8c4-11547f313ca3 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 597.072093] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae93381f-4d00-49b8-b708-6bf1a2420d9e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.081611] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b3b97b6-a851-4e0c-9c31-de40910d6ab2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.099856] env[61868]: DEBUG nova.compute.provider_tree [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 597.121312] env[61868]: DEBUG nova.compute.manager [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 597.127126] env[61868]: DEBUG nova.compute.manager [req-ee8bf827-370f-41f0-894b-b4a7a2a6138c req-19cc036b-9a8e-4eb8-9a98-0f3e4f2708e0 service nova] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Received event network-vif-deleted-d78df532-d2e6-447c-b5d8-9eecf145522c {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 597.212843] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2aaa579b-7f29-4d04-8c99-69eb3c88d449 tempest-ServersAdminNegativeTestJSON-960826170 tempest-ServersAdminNegativeTestJSON-960826170-project-member] Lock "be7273d1-963e-44bf-9862-35703e37081d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.407s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.603586] env[61868]: DEBUG nova.scheduler.client.report [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 597.717204] env[61868]: DEBUG nova.compute.manager [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 598.050676] env[61868]: ERROR nova.compute.manager [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e00fd5a7-bf99-4fb6-a8c4-11547f313ca3, please check neutron logs for more information. [ 598.050676] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 598.050676] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 598.050676] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 598.050676] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 598.050676] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 598.050676] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 598.050676] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 598.050676] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.050676] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 598.050676] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.050676] env[61868]: ERROR nova.compute.manager raise self.value [ 598.050676] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 598.050676] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 598.050676] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.050676] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 598.051632] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.051632] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 598.051632] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e00fd5a7-bf99-4fb6-a8c4-11547f313ca3, please check neutron logs for more information. [ 598.051632] env[61868]: ERROR nova.compute.manager [ 598.051632] env[61868]: Traceback (most recent call last): [ 598.051632] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 598.051632] env[61868]: listener.cb(fileno) [ 598.051632] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 598.051632] env[61868]: result = function(*args, **kwargs) [ 598.051632] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 598.051632] env[61868]: return func(*args, **kwargs) [ 598.051632] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 598.051632] env[61868]: raise e [ 598.051632] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 598.051632] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 598.051632] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 598.051632] env[61868]: created_port_ids = self._update_ports_for_instance( [ 598.051632] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 598.051632] env[61868]: with excutils.save_and_reraise_exception(): [ 598.051632] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.051632] env[61868]: self.force_reraise() [ 598.051632] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.051632] env[61868]: raise self.value [ 598.051632] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 598.051632] env[61868]: updated_port = self._update_port( [ 598.051632] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.051632] env[61868]: _ensure_no_port_binding_failure(port) [ 598.051632] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.051632] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 598.052470] env[61868]: nova.exception.PortBindingFailed: Binding failed for port e00fd5a7-bf99-4fb6-a8c4-11547f313ca3, please check neutron logs for more information. [ 598.052470] env[61868]: Removing descriptor: 16 [ 598.108990] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.004s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 598.109728] env[61868]: ERROR nova.compute.manager [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 88ef3ec5-fe82-4884-80a4-a41aab28f6a7, please check neutron logs for more information. [ 598.109728] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Traceback (most recent call last): [ 598.109728] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 598.109728] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] self.driver.spawn(context, instance, image_meta, [ 598.109728] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 598.109728] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] self._vmops.spawn(context, instance, image_meta, injected_files, [ 598.109728] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 598.109728] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] vm_ref = self.build_virtual_machine(instance, [ 598.109728] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 598.109728] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] vif_infos = vmwarevif.get_vif_info(self._session, [ 598.109728] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 598.110291] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] for vif in network_info: [ 598.110291] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 598.110291] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] return self._sync_wrapper(fn, *args, **kwargs) [ 598.110291] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 598.110291] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] self.wait() [ 598.110291] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 598.110291] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] self[:] = self._gt.wait() [ 598.110291] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 598.110291] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] return self._exit_event.wait() [ 598.110291] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 598.110291] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] result = hub.switch() [ 598.110291] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 598.110291] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] return self.greenlet.switch() [ 598.110839] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 598.110839] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] result = function(*args, **kwargs) [ 598.110839] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 598.110839] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] return func(*args, **kwargs) [ 598.110839] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 598.110839] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] raise e [ 598.110839] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 598.110839] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] nwinfo = self.network_api.allocate_for_instance( [ 598.110839] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 598.110839] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] created_port_ids = self._update_ports_for_instance( [ 598.110839] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 598.110839] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] with excutils.save_and_reraise_exception(): [ 598.110839] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.111849] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] self.force_reraise() [ 598.111849] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.111849] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] raise self.value [ 598.111849] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 598.111849] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] updated_port = self._update_port( [ 598.111849] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.111849] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] _ensure_no_port_binding_failure(port) [ 598.111849] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.111849] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] raise exception.PortBindingFailed(port_id=port['id']) [ 598.111849] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] nova.exception.PortBindingFailed: Binding failed for port 88ef3ec5-fe82-4884-80a4-a41aab28f6a7, please check neutron logs for more information. [ 598.111849] env[61868]: ERROR nova.compute.manager [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] [ 598.112557] env[61868]: DEBUG nova.compute.utils [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Binding failed for port 88ef3ec5-fe82-4884-80a4-a41aab28f6a7, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 598.112557] env[61868]: DEBUG oslo_concurrency.lockutils [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.499s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.113257] env[61868]: INFO nova.compute.claims [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] [instance: e7439426-722a-4420-8607-903cf96dff06] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 598.116375] env[61868]: DEBUG nova.compute.manager [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Build of instance 7965772b-4dd7-4807-8be4-67c6565f9304 was re-scheduled: Binding failed for port 88ef3ec5-fe82-4884-80a4-a41aab28f6a7, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 598.116847] env[61868]: DEBUG nova.compute.manager [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 598.117073] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] Acquiring lock "refresh_cache-7965772b-4dd7-4807-8be4-67c6565f9304" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 598.117218] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] Acquired lock "refresh_cache-7965772b-4dd7-4807-8be4-67c6565f9304" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.117379] env[61868]: DEBUG nova.network.neutron [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 598.133778] env[61868]: DEBUG nova.compute.manager [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 598.172185] env[61868]: DEBUG nova.virt.hardware [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 598.172185] env[61868]: DEBUG nova.virt.hardware [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 598.172185] env[61868]: DEBUG nova.virt.hardware [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 598.172505] env[61868]: DEBUG nova.virt.hardware [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 598.172505] env[61868]: DEBUG nova.virt.hardware [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 598.172505] env[61868]: DEBUG nova.virt.hardware [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 598.172758] env[61868]: DEBUG nova.virt.hardware [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 598.172983] env[61868]: DEBUG nova.virt.hardware [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 598.173169] env[61868]: DEBUG nova.virt.hardware [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 598.173337] env[61868]: DEBUG nova.virt.hardware [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 598.173509] env[61868]: DEBUG nova.virt.hardware [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 598.174489] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19bc2621-d4ec-4ec2-ae14-271a164d4f2d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.184194] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-862cb31a-7dc7-4fa9-8f89-46c5287dd959 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.201408] env[61868]: ERROR nova.compute.manager [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e00fd5a7-bf99-4fb6-a8c4-11547f313ca3, please check neutron logs for more information. [ 598.201408] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] Traceback (most recent call last): [ 598.201408] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 598.201408] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] yield resources [ 598.201408] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 598.201408] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] self.driver.spawn(context, instance, image_meta, [ 598.201408] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 598.201408] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] self._vmops.spawn(context, instance, image_meta, injected_files, [ 598.201408] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 598.201408] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] vm_ref = self.build_virtual_machine(instance, [ 598.201408] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 598.202477] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] vif_infos = vmwarevif.get_vif_info(self._session, [ 598.202477] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 598.202477] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] for vif in network_info: [ 598.202477] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 598.202477] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] return self._sync_wrapper(fn, *args, **kwargs) [ 598.202477] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 598.202477] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] self.wait() [ 598.202477] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 598.202477] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] self[:] = self._gt.wait() [ 598.202477] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 598.202477] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] return self._exit_event.wait() [ 598.202477] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 598.202477] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] current.throw(*self._exc) [ 598.202908] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 598.202908] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] result = function(*args, **kwargs) [ 598.202908] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 598.202908] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] return func(*args, **kwargs) [ 598.202908] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 598.202908] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] raise e [ 598.202908] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 598.202908] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] nwinfo = self.network_api.allocate_for_instance( [ 598.202908] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 598.202908] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] created_port_ids = self._update_ports_for_instance( [ 598.202908] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 598.202908] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] with excutils.save_and_reraise_exception(): [ 598.202908] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.203998] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] self.force_reraise() [ 598.203998] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.203998] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] raise self.value [ 598.203998] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 598.203998] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] updated_port = self._update_port( [ 598.203998] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.203998] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] _ensure_no_port_binding_failure(port) [ 598.203998] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.203998] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] raise exception.PortBindingFailed(port_id=port['id']) [ 598.203998] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] nova.exception.PortBindingFailed: Binding failed for port e00fd5a7-bf99-4fb6-a8c4-11547f313ca3, please check neutron logs for more information. [ 598.203998] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] [ 598.203998] env[61868]: INFO nova.compute.manager [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Terminating instance [ 598.204392] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] Acquiring lock "refresh_cache-5e812408-6f27-4b52-9564-4063f4526340" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 598.204392] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] Acquired lock "refresh_cache-5e812408-6f27-4b52-9564-4063f4526340" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.204630] env[61868]: DEBUG nova.network.neutron [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 598.244885] env[61868]: DEBUG oslo_concurrency.lockutils [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.645054] env[61868]: DEBUG nova.network.neutron [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 598.725605] env[61868]: DEBUG nova.network.neutron [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 598.883491] env[61868]: DEBUG nova.network.neutron [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.928600] env[61868]: DEBUG nova.network.neutron [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.390176] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] Releasing lock "refresh_cache-7965772b-4dd7-4807-8be4-67c6565f9304" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.390465] env[61868]: DEBUG nova.compute.manager [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 599.390784] env[61868]: DEBUG nova.compute.manager [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 599.390784] env[61868]: DEBUG nova.network.neutron [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 599.410326] env[61868]: DEBUG nova.network.neutron [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 599.431162] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] Releasing lock "refresh_cache-5e812408-6f27-4b52-9564-4063f4526340" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.431522] env[61868]: DEBUG nova.compute.manager [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 599.431765] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 599.432044] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4377ec44-228a-43b4-90e2-603430fd956d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.447587] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41dbd398-902a-4dc1-957d-9afb904b7a65 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.473311] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5e812408-6f27-4b52-9564-4063f4526340 could not be found. [ 599.473502] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 599.473692] env[61868]: INFO nova.compute.manager [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Took 0.04 seconds to destroy the instance on the hypervisor. [ 599.473934] env[61868]: DEBUG oslo.service.loopingcall [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 599.476685] env[61868]: DEBUG nova.compute.manager [-] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 599.476790] env[61868]: DEBUG nova.network.neutron [-] [instance: 5e812408-6f27-4b52-9564-4063f4526340] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 599.497168] env[61868]: DEBUG nova.network.neutron [-] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 599.572676] env[61868]: DEBUG nova.compute.manager [req-1d1f6bba-4a24-4cb2-8243-23ddea59d37f req-8587a0c0-9831-43d4-9e8e-09cb8461a246 service nova] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Received event network-changed-e00fd5a7-bf99-4fb6-a8c4-11547f313ca3 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 599.572800] env[61868]: DEBUG nova.compute.manager [req-1d1f6bba-4a24-4cb2-8243-23ddea59d37f req-8587a0c0-9831-43d4-9e8e-09cb8461a246 service nova] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Refreshing instance network info cache due to event network-changed-e00fd5a7-bf99-4fb6-a8c4-11547f313ca3. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 599.572907] env[61868]: DEBUG oslo_concurrency.lockutils [req-1d1f6bba-4a24-4cb2-8243-23ddea59d37f req-8587a0c0-9831-43d4-9e8e-09cb8461a246 service nova] Acquiring lock "refresh_cache-5e812408-6f27-4b52-9564-4063f4526340" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 599.573063] env[61868]: DEBUG oslo_concurrency.lockutils [req-1d1f6bba-4a24-4cb2-8243-23ddea59d37f req-8587a0c0-9831-43d4-9e8e-09cb8461a246 service nova] Acquired lock "refresh_cache-5e812408-6f27-4b52-9564-4063f4526340" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 599.573223] env[61868]: DEBUG nova.network.neutron [req-1d1f6bba-4a24-4cb2-8243-23ddea59d37f req-8587a0c0-9831-43d4-9e8e-09cb8461a246 service nova] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Refreshing network info cache for port e00fd5a7-bf99-4fb6-a8c4-11547f313ca3 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 599.621275] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08b1773c-55e1-49f6-8ec2-b39de6e4b27f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.629056] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c96ea20-e812-4f81-93ef-2e9bed11c5f6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.666271] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e85a6011-9efc-42e3-998f-200accb3336f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.675584] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ef4f3e0-5afe-40c0-ad56-3bad92421aff {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.690729] env[61868]: DEBUG nova.compute.provider_tree [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 599.913500] env[61868]: DEBUG nova.network.neutron [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.000225] env[61868]: DEBUG nova.network.neutron [-] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.093726] env[61868]: DEBUG nova.network.neutron [req-1d1f6bba-4a24-4cb2-8243-23ddea59d37f req-8587a0c0-9831-43d4-9e8e-09cb8461a246 service nova] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 600.162820] env[61868]: DEBUG nova.network.neutron [req-1d1f6bba-4a24-4cb2-8243-23ddea59d37f req-8587a0c0-9831-43d4-9e8e-09cb8461a246 service nova] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.193976] env[61868]: DEBUG nova.scheduler.client.report [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 600.418036] env[61868]: INFO nova.compute.manager [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] [instance: 7965772b-4dd7-4807-8be4-67c6565f9304] Took 1.03 seconds to deallocate network for instance. [ 600.504562] env[61868]: INFO nova.compute.manager [-] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Took 1.03 seconds to deallocate network for instance. [ 600.508534] env[61868]: DEBUG nova.compute.claims [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 600.508734] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.664686] env[61868]: DEBUG oslo_concurrency.lockutils [req-1d1f6bba-4a24-4cb2-8243-23ddea59d37f req-8587a0c0-9831-43d4-9e8e-09cb8461a246 service nova] Releasing lock "refresh_cache-5e812408-6f27-4b52-9564-4063f4526340" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 600.664982] env[61868]: DEBUG nova.compute.manager [req-1d1f6bba-4a24-4cb2-8243-23ddea59d37f req-8587a0c0-9831-43d4-9e8e-09cb8461a246 service nova] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Received event network-vif-deleted-e00fd5a7-bf99-4fb6-a8c4-11547f313ca3 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 600.704395] env[61868]: DEBUG oslo_concurrency.lockutils [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.591s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.705613] env[61868]: DEBUG nova.compute.manager [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] [instance: e7439426-722a-4420-8607-903cf96dff06] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 600.711333] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.005s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.717135] env[61868]: INFO nova.compute.claims [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 601.221513] env[61868]: DEBUG nova.compute.utils [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 601.222920] env[61868]: DEBUG nova.compute.manager [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] [instance: e7439426-722a-4420-8607-903cf96dff06] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 601.223814] env[61868]: DEBUG nova.network.neutron [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] [instance: e7439426-722a-4420-8607-903cf96dff06] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 601.301862] env[61868]: DEBUG nova.policy [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '284c3b21a4154302b3f14396077fa861', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a847996e108f48a9ad85fcf21e86dd61', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 601.459195] env[61868]: INFO nova.scheduler.client.report [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] Deleted allocations for instance 7965772b-4dd7-4807-8be4-67c6565f9304 [ 601.730028] env[61868]: DEBUG nova.compute.manager [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] [instance: e7439426-722a-4420-8607-903cf96dff06] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 601.883308] env[61868]: DEBUG nova.network.neutron [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] [instance: e7439426-722a-4420-8607-903cf96dff06] Successfully created port: af2beab0-e63a-4fd4-b1b8-999a8e0018d7 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 601.970779] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f04ac19c-c09b-4304-93e0-6f82ef1db797 tempest-VolumesAssistedSnapshotsTest-481153157 tempest-VolumesAssistedSnapshotsTest-481153157-project-member] Lock "7965772b-4dd7-4807-8be4-67c6565f9304" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.332s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.312788] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26bd4df0-28f8-4138-8550-6e32aaa537eb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.323375] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8f3a9ec-c38f-49ef-aa8c-03d03bdae95d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.358214] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c54f640-7e2d-49a7-bdf6-a3696258a17d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.371803] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce2de51a-f043-4415-991a-7df06e9afde8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.385463] env[61868]: DEBUG nova.compute.provider_tree [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 602.476045] env[61868]: DEBUG nova.compute.manager [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 602.744721] env[61868]: DEBUG nova.compute.manager [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] [instance: e7439426-722a-4420-8607-903cf96dff06] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 602.788377] env[61868]: DEBUG nova.virt.hardware [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 602.788673] env[61868]: DEBUG nova.virt.hardware [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 602.788990] env[61868]: DEBUG nova.virt.hardware [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 602.789239] env[61868]: DEBUG nova.virt.hardware [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 602.789388] env[61868]: DEBUG nova.virt.hardware [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 602.789616] env[61868]: DEBUG nova.virt.hardware [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 602.789832] env[61868]: DEBUG nova.virt.hardware [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 602.790028] env[61868]: DEBUG nova.virt.hardware [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 602.790178] env[61868]: DEBUG nova.virt.hardware [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 602.790339] env[61868]: DEBUG nova.virt.hardware [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 602.790567] env[61868]: DEBUG nova.virt.hardware [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 602.792073] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-933d250b-35b9-42d7-b6d5-c7e0aad78d83 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.804245] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c4bf0e8-e3b7-4a32-89d9-22f781b04039 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.891325] env[61868]: DEBUG nova.scheduler.client.report [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 603.000047] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.168250] env[61868]: ERROR nova.compute.manager [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port af2beab0-e63a-4fd4-b1b8-999a8e0018d7, please check neutron logs for more information. [ 603.168250] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 603.168250] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 603.168250] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 603.168250] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 603.168250] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 603.168250] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 603.168250] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 603.168250] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 603.168250] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 603.168250] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 603.168250] env[61868]: ERROR nova.compute.manager raise self.value [ 603.168250] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 603.168250] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 603.168250] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 603.168250] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 603.168803] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 603.168803] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 603.168803] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port af2beab0-e63a-4fd4-b1b8-999a8e0018d7, please check neutron logs for more information. [ 603.168803] env[61868]: ERROR nova.compute.manager [ 603.168803] env[61868]: Traceback (most recent call last): [ 603.168803] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 603.168803] env[61868]: listener.cb(fileno) [ 603.168803] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 603.168803] env[61868]: result = function(*args, **kwargs) [ 603.168803] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 603.168803] env[61868]: return func(*args, **kwargs) [ 603.168803] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 603.168803] env[61868]: raise e [ 603.168803] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 603.168803] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 603.168803] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 603.168803] env[61868]: created_port_ids = self._update_ports_for_instance( [ 603.168803] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 603.168803] env[61868]: with excutils.save_and_reraise_exception(): [ 603.168803] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 603.168803] env[61868]: self.force_reraise() [ 603.168803] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 603.168803] env[61868]: raise self.value [ 603.168803] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 603.168803] env[61868]: updated_port = self._update_port( [ 603.168803] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 603.168803] env[61868]: _ensure_no_port_binding_failure(port) [ 603.168803] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 603.168803] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 603.169704] env[61868]: nova.exception.PortBindingFailed: Binding failed for port af2beab0-e63a-4fd4-b1b8-999a8e0018d7, please check neutron logs for more information. [ 603.169704] env[61868]: Removing descriptor: 15 [ 603.169704] env[61868]: ERROR nova.compute.manager [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] [instance: e7439426-722a-4420-8607-903cf96dff06] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port af2beab0-e63a-4fd4-b1b8-999a8e0018d7, please check neutron logs for more information. [ 603.169704] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] Traceback (most recent call last): [ 603.169704] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 603.169704] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] yield resources [ 603.169704] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 603.169704] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] self.driver.spawn(context, instance, image_meta, [ 603.169704] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 603.169704] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] self._vmops.spawn(context, instance, image_meta, injected_files, [ 603.169704] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 603.169704] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] vm_ref = self.build_virtual_machine(instance, [ 603.170080] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 603.170080] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] vif_infos = vmwarevif.get_vif_info(self._session, [ 603.170080] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 603.170080] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] for vif in network_info: [ 603.170080] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 603.170080] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] return self._sync_wrapper(fn, *args, **kwargs) [ 603.170080] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 603.170080] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] self.wait() [ 603.170080] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 603.170080] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] self[:] = self._gt.wait() [ 603.170080] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 603.170080] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] return self._exit_event.wait() [ 603.170080] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 603.170477] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] result = hub.switch() [ 603.170477] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 603.170477] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] return self.greenlet.switch() [ 603.170477] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 603.170477] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] result = function(*args, **kwargs) [ 603.170477] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 603.170477] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] return func(*args, **kwargs) [ 603.170477] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 603.170477] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] raise e [ 603.170477] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 603.170477] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] nwinfo = self.network_api.allocate_for_instance( [ 603.170477] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 603.170477] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] created_port_ids = self._update_ports_for_instance( [ 603.171195] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 603.171195] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] with excutils.save_and_reraise_exception(): [ 603.171195] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 603.171195] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] self.force_reraise() [ 603.171195] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 603.171195] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] raise self.value [ 603.171195] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 603.171195] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] updated_port = self._update_port( [ 603.171195] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 603.171195] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] _ensure_no_port_binding_failure(port) [ 603.171195] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 603.171195] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] raise exception.PortBindingFailed(port_id=port['id']) [ 603.171558] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] nova.exception.PortBindingFailed: Binding failed for port af2beab0-e63a-4fd4-b1b8-999a8e0018d7, please check neutron logs for more information. [ 603.171558] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] [ 603.171558] env[61868]: INFO nova.compute.manager [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] [instance: e7439426-722a-4420-8607-903cf96dff06] Terminating instance [ 603.174872] env[61868]: DEBUG oslo_concurrency.lockutils [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] Acquiring lock "refresh_cache-e7439426-722a-4420-8607-903cf96dff06" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 603.174872] env[61868]: DEBUG oslo_concurrency.lockutils [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] Acquired lock "refresh_cache-e7439426-722a-4420-8607-903cf96dff06" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.175134] env[61868]: DEBUG nova.network.neutron [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] [instance: e7439426-722a-4420-8607-903cf96dff06] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 603.398055] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.687s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 603.398601] env[61868]: DEBUG nova.compute.manager [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 603.406018] env[61868]: DEBUG oslo_concurrency.lockutils [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.638s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.406018] env[61868]: INFO nova.compute.claims [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 603.713273] env[61868]: DEBUG nova.compute.manager [req-c99783e8-8188-4d28-b51f-da9130ea2597 req-40bc2bbb-7a55-4dfa-93fc-d65da167c692 service nova] [instance: e7439426-722a-4420-8607-903cf96dff06] Received event network-changed-af2beab0-e63a-4fd4-b1b8-999a8e0018d7 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 603.713597] env[61868]: DEBUG nova.compute.manager [req-c99783e8-8188-4d28-b51f-da9130ea2597 req-40bc2bbb-7a55-4dfa-93fc-d65da167c692 service nova] [instance: e7439426-722a-4420-8607-903cf96dff06] Refreshing instance network info cache due to event network-changed-af2beab0-e63a-4fd4-b1b8-999a8e0018d7. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 603.713641] env[61868]: DEBUG oslo_concurrency.lockutils [req-c99783e8-8188-4d28-b51f-da9130ea2597 req-40bc2bbb-7a55-4dfa-93fc-d65da167c692 service nova] Acquiring lock "refresh_cache-e7439426-722a-4420-8607-903cf96dff06" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 603.724789] env[61868]: DEBUG nova.network.neutron [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] [instance: e7439426-722a-4420-8607-903cf96dff06] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 603.910462] env[61868]: DEBUG nova.compute.utils [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 603.913172] env[61868]: DEBUG nova.compute.manager [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 603.913281] env[61868]: DEBUG nova.network.neutron [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 603.953289] env[61868]: DEBUG nova.network.neutron [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] [instance: e7439426-722a-4420-8607-903cf96dff06] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.994742] env[61868]: DEBUG nova.policy [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4013f992abd14226b51fc442f30e90f6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8e0bc864ba72472eb0aab291404e916b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 604.419719] env[61868]: DEBUG nova.compute.manager [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 604.456306] env[61868]: DEBUG oslo_concurrency.lockutils [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] Releasing lock "refresh_cache-e7439426-722a-4420-8607-903cf96dff06" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 604.456793] env[61868]: DEBUG nova.compute.manager [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] [instance: e7439426-722a-4420-8607-903cf96dff06] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 604.457062] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] [instance: e7439426-722a-4420-8607-903cf96dff06] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 604.457902] env[61868]: DEBUG nova.network.neutron [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Successfully created port: f0536e29-ba69-4dc1-a336-c44893ce3afc {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 604.461371] env[61868]: DEBUG oslo_concurrency.lockutils [req-c99783e8-8188-4d28-b51f-da9130ea2597 req-40bc2bbb-7a55-4dfa-93fc-d65da167c692 service nova] Acquired lock "refresh_cache-e7439426-722a-4420-8607-903cf96dff06" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.461586] env[61868]: DEBUG nova.network.neutron [req-c99783e8-8188-4d28-b51f-da9130ea2597 req-40bc2bbb-7a55-4dfa-93fc-d65da167c692 service nova] [instance: e7439426-722a-4420-8607-903cf96dff06] Refreshing network info cache for port af2beab0-e63a-4fd4-b1b8-999a8e0018d7 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 604.463855] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-18986fc6-64a3-4e26-abd1-fb48c5c443a7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.480566] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b3ec22f-8dd6-4cd0-98a8-092c65c15f33 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.510883] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] [instance: e7439426-722a-4420-8607-903cf96dff06] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e7439426-722a-4420-8607-903cf96dff06 could not be found. [ 604.510883] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] [instance: e7439426-722a-4420-8607-903cf96dff06] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 604.510883] env[61868]: INFO nova.compute.manager [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] [instance: e7439426-722a-4420-8607-903cf96dff06] Took 0.05 seconds to destroy the instance on the hypervisor. [ 604.510883] env[61868]: DEBUG oslo.service.loopingcall [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 604.513819] env[61868]: DEBUG nova.compute.manager [-] [instance: e7439426-722a-4420-8607-903cf96dff06] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 604.513918] env[61868]: DEBUG nova.network.neutron [-] [instance: e7439426-722a-4420-8607-903cf96dff06] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 604.550387] env[61868]: DEBUG nova.network.neutron [-] [instance: e7439426-722a-4420-8607-903cf96dff06] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 604.945017] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74f5c1e6-f602-4b4d-b48d-22be7dcd4a6d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.953983] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdd740eb-6c50-4530-abc1-bfadfadd904b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.990958] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e0597ab-168c-473a-99aa-aba0bd553034 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.000860] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5353a11f-bc54-4b0c-a866-ab567d1a15c1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.017046] env[61868]: DEBUG nova.compute.provider_tree [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 605.026978] env[61868]: DEBUG nova.network.neutron [req-c99783e8-8188-4d28-b51f-da9130ea2597 req-40bc2bbb-7a55-4dfa-93fc-d65da167c692 service nova] [instance: e7439426-722a-4420-8607-903cf96dff06] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 605.052647] env[61868]: DEBUG nova.network.neutron [-] [instance: e7439426-722a-4420-8607-903cf96dff06] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.352181] env[61868]: DEBUG nova.network.neutron [req-c99783e8-8188-4d28-b51f-da9130ea2597 req-40bc2bbb-7a55-4dfa-93fc-d65da167c692 service nova] [instance: e7439426-722a-4420-8607-903cf96dff06] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.435106] env[61868]: DEBUG nova.compute.manager [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 605.472834] env[61868]: DEBUG nova.virt.hardware [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 605.473252] env[61868]: DEBUG nova.virt.hardware [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 605.473439] env[61868]: DEBUG nova.virt.hardware [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 605.473609] env[61868]: DEBUG nova.virt.hardware [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 605.473756] env[61868]: DEBUG nova.virt.hardware [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 605.473914] env[61868]: DEBUG nova.virt.hardware [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 605.474117] env[61868]: DEBUG nova.virt.hardware [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 605.474276] env[61868]: DEBUG nova.virt.hardware [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 605.474921] env[61868]: DEBUG nova.virt.hardware [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 605.474921] env[61868]: DEBUG nova.virt.hardware [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 605.474921] env[61868]: DEBUG nova.virt.hardware [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 605.476176] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1175976-0e0a-43e6-8d10-8395caff779b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.491552] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a48f00d9-d67e-4481-93d8-9d7ea5ad491c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.521588] env[61868]: DEBUG nova.scheduler.client.report [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 605.556405] env[61868]: INFO nova.compute.manager [-] [instance: e7439426-722a-4420-8607-903cf96dff06] Took 1.04 seconds to deallocate network for instance. [ 605.557836] env[61868]: DEBUG nova.compute.claims [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] [instance: e7439426-722a-4420-8607-903cf96dff06] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 605.559631] env[61868]: DEBUG oslo_concurrency.lockutils [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.855168] env[61868]: DEBUG oslo_concurrency.lockutils [req-c99783e8-8188-4d28-b51f-da9130ea2597 req-40bc2bbb-7a55-4dfa-93fc-d65da167c692 service nova] Releasing lock "refresh_cache-e7439426-722a-4420-8607-903cf96dff06" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.029594] env[61868]: DEBUG oslo_concurrency.lockutils [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.628s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.030228] env[61868]: DEBUG nova.compute.manager [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 606.033651] env[61868]: DEBUG oslo_concurrency.lockutils [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.518s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.039442] env[61868]: INFO nova.compute.claims [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 606.109234] env[61868]: ERROR nova.compute.manager [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f0536e29-ba69-4dc1-a336-c44893ce3afc, please check neutron logs for more information. [ 606.109234] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 606.109234] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.109234] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 606.109234] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 606.109234] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 606.109234] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 606.109234] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 606.109234] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.109234] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 606.109234] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.109234] env[61868]: ERROR nova.compute.manager raise self.value [ 606.109234] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 606.109234] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 606.109234] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.109234] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 606.110046] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.110046] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 606.110046] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f0536e29-ba69-4dc1-a336-c44893ce3afc, please check neutron logs for more information. [ 606.110046] env[61868]: ERROR nova.compute.manager [ 606.110046] env[61868]: Traceback (most recent call last): [ 606.110046] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 606.110046] env[61868]: listener.cb(fileno) [ 606.110046] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 606.110046] env[61868]: result = function(*args, **kwargs) [ 606.110046] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 606.110046] env[61868]: return func(*args, **kwargs) [ 606.110046] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 606.110046] env[61868]: raise e [ 606.110046] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.110046] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 606.110046] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 606.110046] env[61868]: created_port_ids = self._update_ports_for_instance( [ 606.110046] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 606.110046] env[61868]: with excutils.save_and_reraise_exception(): [ 606.110046] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.110046] env[61868]: self.force_reraise() [ 606.110046] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.110046] env[61868]: raise self.value [ 606.110046] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 606.110046] env[61868]: updated_port = self._update_port( [ 606.110046] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.110046] env[61868]: _ensure_no_port_binding_failure(port) [ 606.110046] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.110046] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 606.111304] env[61868]: nova.exception.PortBindingFailed: Binding failed for port f0536e29-ba69-4dc1-a336-c44893ce3afc, please check neutron logs for more information. [ 606.111304] env[61868]: Removing descriptor: 16 [ 606.111304] env[61868]: ERROR nova.compute.manager [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f0536e29-ba69-4dc1-a336-c44893ce3afc, please check neutron logs for more information. [ 606.111304] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Traceback (most recent call last): [ 606.111304] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 606.111304] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] yield resources [ 606.111304] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 606.111304] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] self.driver.spawn(context, instance, image_meta, [ 606.111304] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 606.111304] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] self._vmops.spawn(context, instance, image_meta, injected_files, [ 606.111304] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 606.111304] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] vm_ref = self.build_virtual_machine(instance, [ 606.113075] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 606.113075] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] vif_infos = vmwarevif.get_vif_info(self._session, [ 606.113075] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 606.113075] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] for vif in network_info: [ 606.113075] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 606.113075] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] return self._sync_wrapper(fn, *args, **kwargs) [ 606.113075] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 606.113075] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] self.wait() [ 606.113075] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 606.113075] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] self[:] = self._gt.wait() [ 606.113075] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 606.113075] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] return self._exit_event.wait() [ 606.113075] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 606.113654] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] result = hub.switch() [ 606.113654] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 606.113654] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] return self.greenlet.switch() [ 606.113654] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 606.113654] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] result = function(*args, **kwargs) [ 606.113654] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 606.113654] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] return func(*args, **kwargs) [ 606.113654] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 606.113654] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] raise e [ 606.113654] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.113654] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] nwinfo = self.network_api.allocate_for_instance( [ 606.113654] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 606.113654] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] created_port_ids = self._update_ports_for_instance( [ 606.114223] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 606.114223] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] with excutils.save_and_reraise_exception(): [ 606.114223] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.114223] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] self.force_reraise() [ 606.114223] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.114223] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] raise self.value [ 606.114223] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 606.114223] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] updated_port = self._update_port( [ 606.114223] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.114223] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] _ensure_no_port_binding_failure(port) [ 606.114223] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.114223] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] raise exception.PortBindingFailed(port_id=port['id']) [ 606.114752] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] nova.exception.PortBindingFailed: Binding failed for port f0536e29-ba69-4dc1-a336-c44893ce3afc, please check neutron logs for more information. [ 606.114752] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] [ 606.114752] env[61868]: INFO nova.compute.manager [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Terminating instance [ 606.114752] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] Acquiring lock "refresh_cache-a93e681e-1cb1-4025-8725-4c6664d78c96" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 606.114752] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] Acquired lock "refresh_cache-a93e681e-1cb1-4025-8725-4c6664d78c96" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.114752] env[61868]: DEBUG nova.network.neutron [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 606.407305] env[61868]: DEBUG nova.compute.manager [req-cde70a47-55dd-4f2d-8470-b4f51fbd2c0e req-6e4a698f-dd49-495c-9846-4749d0a39950 service nova] [instance: e7439426-722a-4420-8607-903cf96dff06] Received event network-vif-deleted-af2beab0-e63a-4fd4-b1b8-999a8e0018d7 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 606.545477] env[61868]: DEBUG nova.compute.utils [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 606.548781] env[61868]: DEBUG nova.compute.manager [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 606.551102] env[61868]: DEBUG nova.network.neutron [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 606.624638] env[61868]: DEBUG nova.policy [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '103e8440b5a04afbbff6d5f049403835', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '84b3958843f040969c2f1955ff403454', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 606.643730] env[61868]: DEBUG nova.network.neutron [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.818548] env[61868]: DEBUG nova.network.neutron [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.049482] env[61868]: DEBUG nova.compute.manager [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 607.200125] env[61868]: DEBUG nova.network.neutron [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Successfully created port: 0d931638-71c7-4f34-aff9-273312a018e6 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 607.322749] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] Releasing lock "refresh_cache-a93e681e-1cb1-4025-8725-4c6664d78c96" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 607.323215] env[61868]: DEBUG nova.compute.manager [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 607.323467] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 607.323955] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7d7c59a6-efdf-4057-8136-bd2108467aa1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.335102] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ac7a486-dfb6-4a6f-89f9-64418d59de11 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.369783] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a93e681e-1cb1-4025-8725-4c6664d78c96 could not be found. [ 607.372029] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 607.372029] env[61868]: INFO nova.compute.manager [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Took 0.05 seconds to destroy the instance on the hypervisor. [ 607.372029] env[61868]: DEBUG oslo.service.loopingcall [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 607.375881] env[61868]: DEBUG nova.compute.manager [-] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 607.375881] env[61868]: DEBUG nova.network.neutron [-] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 607.404496] env[61868]: DEBUG nova.network.neutron [-] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 607.563050] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82045ebd-d00e-48f7-87b8-5935459f777b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.574238] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80c2bb72-1bd0-4e28-90b0-917b03bf8567 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.612659] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba9285b4-023b-4633-bf84-e7df43b2636b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.621422] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0769b14e-74c9-4fe1-af3d-cae056c3c010 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.635754] env[61868]: DEBUG nova.compute.provider_tree [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 607.907106] env[61868]: DEBUG nova.network.neutron [-] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.062598] env[61868]: DEBUG nova.compute.manager [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 608.101709] env[61868]: DEBUG nova.virt.hardware [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:19:17Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1464753298',id=22,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-660449824',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 608.101987] env[61868]: DEBUG nova.virt.hardware [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 608.102165] env[61868]: DEBUG nova.virt.hardware [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 608.102347] env[61868]: DEBUG nova.virt.hardware [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 608.102511] env[61868]: DEBUG nova.virt.hardware [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 608.102629] env[61868]: DEBUG nova.virt.hardware [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 608.102940] env[61868]: DEBUG nova.virt.hardware [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 608.103026] env[61868]: DEBUG nova.virt.hardware [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 608.107419] env[61868]: DEBUG nova.virt.hardware [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 608.107489] env[61868]: DEBUG nova.virt.hardware [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 608.107648] env[61868]: DEBUG nova.virt.hardware [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 608.108556] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-770f75ce-2c1e-41c4-9835-64414552a788 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.122661] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2c149de-c0c5-4a7c-8847-d10cef91d0d9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.141785] env[61868]: DEBUG nova.scheduler.client.report [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 608.411973] env[61868]: INFO nova.compute.manager [-] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Took 1.03 seconds to deallocate network for instance. [ 608.412964] env[61868]: DEBUG nova.compute.claims [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 608.413332] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.651018] env[61868]: DEBUG oslo_concurrency.lockutils [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.614s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.651018] env[61868]: DEBUG nova.compute.manager [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 608.651806] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.105s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.653131] env[61868]: INFO nova.compute.claims [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 608.736881] env[61868]: DEBUG nova.compute.manager [req-0b852162-e7ca-479e-87c1-ee50fbc7059f req-8973c1a3-7abf-406a-a8b4-52e3667017b3 service nova] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Received event network-changed-f0536e29-ba69-4dc1-a336-c44893ce3afc {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 608.736881] env[61868]: DEBUG nova.compute.manager [req-0b852162-e7ca-479e-87c1-ee50fbc7059f req-8973c1a3-7abf-406a-a8b4-52e3667017b3 service nova] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Refreshing instance network info cache due to event network-changed-f0536e29-ba69-4dc1-a336-c44893ce3afc. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 608.736881] env[61868]: DEBUG oslo_concurrency.lockutils [req-0b852162-e7ca-479e-87c1-ee50fbc7059f req-8973c1a3-7abf-406a-a8b4-52e3667017b3 service nova] Acquiring lock "refresh_cache-a93e681e-1cb1-4025-8725-4c6664d78c96" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 608.736881] env[61868]: DEBUG oslo_concurrency.lockutils [req-0b852162-e7ca-479e-87c1-ee50fbc7059f req-8973c1a3-7abf-406a-a8b4-52e3667017b3 service nova] Acquired lock "refresh_cache-a93e681e-1cb1-4025-8725-4c6664d78c96" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.736881] env[61868]: DEBUG nova.network.neutron [req-0b852162-e7ca-479e-87c1-ee50fbc7059f req-8973c1a3-7abf-406a-a8b4-52e3667017b3 service nova] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Refreshing network info cache for port f0536e29-ba69-4dc1-a336-c44893ce3afc {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 609.114112] env[61868]: ERROR nova.compute.manager [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0d931638-71c7-4f34-aff9-273312a018e6, please check neutron logs for more information. [ 609.114112] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 609.114112] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.114112] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 609.114112] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 609.114112] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 609.114112] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 609.114112] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 609.114112] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.114112] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 609.114112] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.114112] env[61868]: ERROR nova.compute.manager raise self.value [ 609.114112] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 609.114112] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 609.114112] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.114112] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 609.114930] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.114930] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 609.114930] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0d931638-71c7-4f34-aff9-273312a018e6, please check neutron logs for more information. [ 609.114930] env[61868]: ERROR nova.compute.manager [ 609.114930] env[61868]: Traceback (most recent call last): [ 609.114930] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 609.114930] env[61868]: listener.cb(fileno) [ 609.114930] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 609.114930] env[61868]: result = function(*args, **kwargs) [ 609.114930] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 609.114930] env[61868]: return func(*args, **kwargs) [ 609.114930] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 609.114930] env[61868]: raise e [ 609.114930] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.114930] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 609.114930] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 609.114930] env[61868]: created_port_ids = self._update_ports_for_instance( [ 609.114930] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 609.114930] env[61868]: with excutils.save_and_reraise_exception(): [ 609.114930] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.114930] env[61868]: self.force_reraise() [ 609.114930] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.114930] env[61868]: raise self.value [ 609.114930] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 609.114930] env[61868]: updated_port = self._update_port( [ 609.114930] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.114930] env[61868]: _ensure_no_port_binding_failure(port) [ 609.114930] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.114930] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 609.115906] env[61868]: nova.exception.PortBindingFailed: Binding failed for port 0d931638-71c7-4f34-aff9-273312a018e6, please check neutron logs for more information. [ 609.115906] env[61868]: Removing descriptor: 16 [ 609.115906] env[61868]: ERROR nova.compute.manager [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0d931638-71c7-4f34-aff9-273312a018e6, please check neutron logs for more information. [ 609.115906] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Traceback (most recent call last): [ 609.115906] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 609.115906] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] yield resources [ 609.115906] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 609.115906] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] self.driver.spawn(context, instance, image_meta, [ 609.115906] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 609.115906] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] self._vmops.spawn(context, instance, image_meta, injected_files, [ 609.115906] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 609.115906] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] vm_ref = self.build_virtual_machine(instance, [ 609.116342] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 609.116342] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] vif_infos = vmwarevif.get_vif_info(self._session, [ 609.116342] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 609.116342] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] for vif in network_info: [ 609.116342] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 609.116342] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] return self._sync_wrapper(fn, *args, **kwargs) [ 609.116342] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 609.116342] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] self.wait() [ 609.116342] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 609.116342] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] self[:] = self._gt.wait() [ 609.116342] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 609.116342] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] return self._exit_event.wait() [ 609.116342] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 609.116783] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] result = hub.switch() [ 609.116783] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 609.116783] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] return self.greenlet.switch() [ 609.116783] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 609.116783] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] result = function(*args, **kwargs) [ 609.116783] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 609.116783] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] return func(*args, **kwargs) [ 609.116783] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 609.116783] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] raise e [ 609.116783] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.116783] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] nwinfo = self.network_api.allocate_for_instance( [ 609.116783] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 609.116783] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] created_port_ids = self._update_ports_for_instance( [ 609.117209] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 609.117209] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] with excutils.save_and_reraise_exception(): [ 609.117209] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.117209] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] self.force_reraise() [ 609.117209] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.117209] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] raise self.value [ 609.117209] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 609.117209] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] updated_port = self._update_port( [ 609.117209] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.117209] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] _ensure_no_port_binding_failure(port) [ 609.117209] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.117209] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] raise exception.PortBindingFailed(port_id=port['id']) [ 609.117608] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] nova.exception.PortBindingFailed: Binding failed for port 0d931638-71c7-4f34-aff9-273312a018e6, please check neutron logs for more information. [ 609.117608] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] [ 609.117608] env[61868]: INFO nova.compute.manager [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Terminating instance [ 609.121500] env[61868]: DEBUG oslo_concurrency.lockutils [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] Acquiring lock "refresh_cache-448b8246-8d41-4154-bb99-b3c66d8e9475" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.121500] env[61868]: DEBUG oslo_concurrency.lockutils [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] Acquired lock "refresh_cache-448b8246-8d41-4154-bb99-b3c66d8e9475" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.121500] env[61868]: DEBUG nova.network.neutron [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 609.157902] env[61868]: DEBUG nova.compute.utils [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 609.167180] env[61868]: DEBUG nova.compute.manager [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 609.167180] env[61868]: DEBUG nova.network.neutron [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 609.232417] env[61868]: DEBUG nova.policy [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0e0af3737bc3469a8d2a05617a731d35', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '942f23f7264c45d0b3ba25fc2398f90d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 609.282061] env[61868]: DEBUG nova.network.neutron [req-0b852162-e7ca-479e-87c1-ee50fbc7059f req-8973c1a3-7abf-406a-a8b4-52e3667017b3 service nova] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 609.458043] env[61868]: DEBUG nova.network.neutron [req-0b852162-e7ca-479e-87c1-ee50fbc7059f req-8973c1a3-7abf-406a-a8b4-52e3667017b3 service nova] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.651070] env[61868]: DEBUG nova.network.neutron [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 609.672469] env[61868]: DEBUG nova.compute.manager [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 609.839849] env[61868]: DEBUG nova.network.neutron [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.891094] env[61868]: DEBUG nova.network.neutron [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Successfully created port: 8354b128-a713-44f1-9205-d5dc691119b0 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 609.960719] env[61868]: DEBUG oslo_concurrency.lockutils [req-0b852162-e7ca-479e-87c1-ee50fbc7059f req-8973c1a3-7abf-406a-a8b4-52e3667017b3 service nova] Releasing lock "refresh_cache-a93e681e-1cb1-4025-8725-4c6664d78c96" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 609.960719] env[61868]: DEBUG nova.compute.manager [req-0b852162-e7ca-479e-87c1-ee50fbc7059f req-8973c1a3-7abf-406a-a8b4-52e3667017b3 service nova] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Received event network-vif-deleted-f0536e29-ba69-4dc1-a336-c44893ce3afc {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 610.127619] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-822d2100-65b1-4ce0-a7ee-05b546b979f7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.139362] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2051387-df95-4544-9234-e9631e04be79 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.174885] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bd87764-a47a-425a-9211-2934ac758412 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.186390] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2de8178-4558-4282-955c-5b6fd5ebe89c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.201389] env[61868]: DEBUG nova.compute.provider_tree [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 610.342658] env[61868]: DEBUG oslo_concurrency.lockutils [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] Releasing lock "refresh_cache-448b8246-8d41-4154-bb99-b3c66d8e9475" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 610.343119] env[61868]: DEBUG nova.compute.manager [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 610.343311] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 610.343600] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c5ff73b9-2d0e-480c-906c-3b57b97576f2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.354120] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8b9fc1f-8bd2-4480-94a6-9a5226e8979c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.378934] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 448b8246-8d41-4154-bb99-b3c66d8e9475 could not be found. [ 610.379233] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 610.379511] env[61868]: INFO nova.compute.manager [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Took 0.04 seconds to destroy the instance on the hypervisor. [ 610.379669] env[61868]: DEBUG oslo.service.loopingcall [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 610.379910] env[61868]: DEBUG nova.compute.manager [-] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 610.380050] env[61868]: DEBUG nova.network.neutron [-] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 610.404104] env[61868]: DEBUG nova.network.neutron [-] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 610.690305] env[61868]: DEBUG nova.compute.manager [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 610.704983] env[61868]: DEBUG nova.scheduler.client.report [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 610.735328] env[61868]: DEBUG nova.virt.hardware [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 610.735602] env[61868]: DEBUG nova.virt.hardware [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 610.735761] env[61868]: DEBUG nova.virt.hardware [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 610.735956] env[61868]: DEBUG nova.virt.hardware [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 610.736129] env[61868]: DEBUG nova.virt.hardware [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 610.736285] env[61868]: DEBUG nova.virt.hardware [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 610.736467] env[61868]: DEBUG nova.virt.hardware [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 610.736815] env[61868]: DEBUG nova.virt.hardware [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 610.736815] env[61868]: DEBUG nova.virt.hardware [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 610.736941] env[61868]: DEBUG nova.virt.hardware [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 610.737115] env[61868]: DEBUG nova.virt.hardware [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 610.738205] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f98d4d2c-de38-4e51-a69d-cf157101d5a6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.747337] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03b1bd2c-9fa6-4675-b313-0b1b40e0a0b2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.908528] env[61868]: DEBUG nova.network.neutron [-] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.063832] env[61868]: DEBUG nova.compute.manager [req-8b021203-8a76-4766-b4a4-31d32b728fbf req-c644e67f-ce76-448d-aded-6b4ed8115908 service nova] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Received event network-changed-0d931638-71c7-4f34-aff9-273312a018e6 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 611.064037] env[61868]: DEBUG nova.compute.manager [req-8b021203-8a76-4766-b4a4-31d32b728fbf req-c644e67f-ce76-448d-aded-6b4ed8115908 service nova] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Refreshing instance network info cache due to event network-changed-0d931638-71c7-4f34-aff9-273312a018e6. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 611.064253] env[61868]: DEBUG oslo_concurrency.lockutils [req-8b021203-8a76-4766-b4a4-31d32b728fbf req-c644e67f-ce76-448d-aded-6b4ed8115908 service nova] Acquiring lock "refresh_cache-448b8246-8d41-4154-bb99-b3c66d8e9475" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.064403] env[61868]: DEBUG oslo_concurrency.lockutils [req-8b021203-8a76-4766-b4a4-31d32b728fbf req-c644e67f-ce76-448d-aded-6b4ed8115908 service nova] Acquired lock "refresh_cache-448b8246-8d41-4154-bb99-b3c66d8e9475" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.064547] env[61868]: DEBUG nova.network.neutron [req-8b021203-8a76-4766-b4a4-31d32b728fbf req-c644e67f-ce76-448d-aded-6b4ed8115908 service nova] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Refreshing network info cache for port 0d931638-71c7-4f34-aff9-273312a018e6 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 611.210345] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.558s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 611.210879] env[61868]: DEBUG nova.compute.manager [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 611.219021] env[61868]: DEBUG oslo_concurrency.lockutils [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.511s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.219021] env[61868]: DEBUG nova.objects.instance [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Lazy-loading 'resources' on Instance uuid ddde9ee7-5198-42e5-86b9-f51727dfbf60 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 611.410198] env[61868]: INFO nova.compute.manager [-] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Took 1.03 seconds to deallocate network for instance. [ 611.412915] env[61868]: DEBUG nova.compute.claims [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 611.413120] env[61868]: DEBUG oslo_concurrency.lockutils [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.572684] env[61868]: ERROR nova.compute.manager [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8354b128-a713-44f1-9205-d5dc691119b0, please check neutron logs for more information. [ 611.572684] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 611.572684] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.572684] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 611.572684] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.572684] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 611.572684] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.572684] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 611.572684] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.572684] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 611.572684] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.572684] env[61868]: ERROR nova.compute.manager raise self.value [ 611.572684] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.572684] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 611.572684] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.572684] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 611.573538] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.573538] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 611.573538] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8354b128-a713-44f1-9205-d5dc691119b0, please check neutron logs for more information. [ 611.573538] env[61868]: ERROR nova.compute.manager [ 611.573538] env[61868]: Traceback (most recent call last): [ 611.573538] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 611.573538] env[61868]: listener.cb(fileno) [ 611.573538] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.573538] env[61868]: result = function(*args, **kwargs) [ 611.573538] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 611.573538] env[61868]: return func(*args, **kwargs) [ 611.573538] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 611.573538] env[61868]: raise e [ 611.573538] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.573538] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 611.573538] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.573538] env[61868]: created_port_ids = self._update_ports_for_instance( [ 611.573538] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.573538] env[61868]: with excutils.save_and_reraise_exception(): [ 611.573538] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.573538] env[61868]: self.force_reraise() [ 611.573538] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.573538] env[61868]: raise self.value [ 611.573538] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.573538] env[61868]: updated_port = self._update_port( [ 611.573538] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.573538] env[61868]: _ensure_no_port_binding_failure(port) [ 611.573538] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.573538] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 611.574897] env[61868]: nova.exception.PortBindingFailed: Binding failed for port 8354b128-a713-44f1-9205-d5dc691119b0, please check neutron logs for more information. [ 611.574897] env[61868]: Removing descriptor: 16 [ 611.574897] env[61868]: ERROR nova.compute.manager [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8354b128-a713-44f1-9205-d5dc691119b0, please check neutron logs for more information. [ 611.574897] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Traceback (most recent call last): [ 611.574897] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 611.574897] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] yield resources [ 611.574897] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 611.574897] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] self.driver.spawn(context, instance, image_meta, [ 611.574897] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 611.574897] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 611.574897] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 611.574897] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] vm_ref = self.build_virtual_machine(instance, [ 611.575560] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 611.575560] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] vif_infos = vmwarevif.get_vif_info(self._session, [ 611.575560] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 611.575560] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] for vif in network_info: [ 611.575560] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 611.575560] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] return self._sync_wrapper(fn, *args, **kwargs) [ 611.575560] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 611.575560] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] self.wait() [ 611.575560] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 611.575560] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] self[:] = self._gt.wait() [ 611.575560] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 611.575560] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] return self._exit_event.wait() [ 611.575560] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 611.576371] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] result = hub.switch() [ 611.576371] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 611.576371] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] return self.greenlet.switch() [ 611.576371] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.576371] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] result = function(*args, **kwargs) [ 611.576371] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 611.576371] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] return func(*args, **kwargs) [ 611.576371] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 611.576371] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] raise e [ 611.576371] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.576371] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] nwinfo = self.network_api.allocate_for_instance( [ 611.576371] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.576371] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] created_port_ids = self._update_ports_for_instance( [ 611.577884] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.577884] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] with excutils.save_and_reraise_exception(): [ 611.577884] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.577884] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] self.force_reraise() [ 611.577884] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.577884] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] raise self.value [ 611.577884] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.577884] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] updated_port = self._update_port( [ 611.577884] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.577884] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] _ensure_no_port_binding_failure(port) [ 611.577884] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.577884] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] raise exception.PortBindingFailed(port_id=port['id']) [ 611.582358] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] nova.exception.PortBindingFailed: Binding failed for port 8354b128-a713-44f1-9205-d5dc691119b0, please check neutron logs for more information. [ 611.582358] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] [ 611.582358] env[61868]: INFO nova.compute.manager [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Terminating instance [ 611.582358] env[61868]: DEBUG oslo_concurrency.lockutils [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] Acquiring lock "refresh_cache-2f9deec0-e0ad-4cb4-9868-9aa53cdf602d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.582358] env[61868]: DEBUG oslo_concurrency.lockutils [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] Acquired lock "refresh_cache-2f9deec0-e0ad-4cb4-9868-9aa53cdf602d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.582358] env[61868]: DEBUG nova.network.neutron [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 611.590767] env[61868]: DEBUG nova.network.neutron [req-8b021203-8a76-4766-b4a4-31d32b728fbf req-c644e67f-ce76-448d-aded-6b4ed8115908 service nova] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.719990] env[61868]: DEBUG nova.compute.utils [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 611.726024] env[61868]: DEBUG nova.compute.manager [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Not allocating networking since 'none' was specified. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 611.752735] env[61868]: DEBUG nova.network.neutron [req-8b021203-8a76-4766-b4a4-31d32b728fbf req-c644e67f-ce76-448d-aded-6b4ed8115908 service nova] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.104632] env[61868]: DEBUG nova.network.neutron [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.224696] env[61868]: DEBUG nova.compute.manager [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 612.234629] env[61868]: DEBUG nova.network.neutron [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.241209] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] Acquiring lock "66c5f446-91e5-4de1-abb6-f06a0a36020d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.241481] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] Lock "66c5f446-91e5-4de1-abb6-f06a0a36020d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.249769] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55a1e6ac-7bf2-4a9f-a264-57ee7af070f0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.255599] env[61868]: DEBUG oslo_concurrency.lockutils [req-8b021203-8a76-4766-b4a4-31d32b728fbf req-c644e67f-ce76-448d-aded-6b4ed8115908 service nova] Releasing lock "refresh_cache-448b8246-8d41-4154-bb99-b3c66d8e9475" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.255856] env[61868]: DEBUG nova.compute.manager [req-8b021203-8a76-4766-b4a4-31d32b728fbf req-c644e67f-ce76-448d-aded-6b4ed8115908 service nova] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Received event network-vif-deleted-0d931638-71c7-4f34-aff9-273312a018e6 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 612.261607] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3888111-0674-474a-8a77-4bef116303a9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.308493] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b913975-4416-4ac0-a06a-5e75ac6cf60d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.317544] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d75037b-dd8d-4544-aebb-9f7e2e1bd30e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.333421] env[61868]: DEBUG nova.compute.provider_tree [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 612.742262] env[61868]: DEBUG oslo_concurrency.lockutils [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] Releasing lock "refresh_cache-2f9deec0-e0ad-4cb4-9868-9aa53cdf602d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.742874] env[61868]: DEBUG nova.compute.manager [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 612.742966] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 612.743209] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e13d0950-0cee-48a2-9451-b3f45745abe8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.756632] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-484b32af-6cd5-4693-bb92-09c54fd53c3d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.781296] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d could not be found. [ 612.781603] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 612.781800] env[61868]: INFO nova.compute.manager [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Took 0.04 seconds to destroy the instance on the hypervisor. [ 612.782188] env[61868]: DEBUG oslo.service.loopingcall [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 612.782447] env[61868]: DEBUG nova.compute.manager [-] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 612.782593] env[61868]: DEBUG nova.network.neutron [-] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 612.808383] env[61868]: DEBUG nova.network.neutron [-] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.836331] env[61868]: DEBUG nova.scheduler.client.report [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 613.236102] env[61868]: DEBUG nova.compute.manager [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 613.266234] env[61868]: DEBUG nova.compute.manager [req-045d4c8f-b353-4c0a-a828-0e21af46a6c9 req-6576c2f0-3ca9-4699-9f62-a78af2a1f644 service nova] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Received event network-changed-8354b128-a713-44f1-9205-d5dc691119b0 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 613.266234] env[61868]: DEBUG nova.compute.manager [req-045d4c8f-b353-4c0a-a828-0e21af46a6c9 req-6576c2f0-3ca9-4699-9f62-a78af2a1f644 service nova] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Refreshing instance network info cache due to event network-changed-8354b128-a713-44f1-9205-d5dc691119b0. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 613.266234] env[61868]: DEBUG oslo_concurrency.lockutils [req-045d4c8f-b353-4c0a-a828-0e21af46a6c9 req-6576c2f0-3ca9-4699-9f62-a78af2a1f644 service nova] Acquiring lock "refresh_cache-2f9deec0-e0ad-4cb4-9868-9aa53cdf602d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 613.267062] env[61868]: DEBUG oslo_concurrency.lockutils [req-045d4c8f-b353-4c0a-a828-0e21af46a6c9 req-6576c2f0-3ca9-4699-9f62-a78af2a1f644 service nova] Acquired lock "refresh_cache-2f9deec0-e0ad-4cb4-9868-9aa53cdf602d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.267263] env[61868]: DEBUG nova.network.neutron [req-045d4c8f-b353-4c0a-a828-0e21af46a6c9 req-6576c2f0-3ca9-4699-9f62-a78af2a1f644 service nova] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Refreshing network info cache for port 8354b128-a713-44f1-9205-d5dc691119b0 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 613.271707] env[61868]: DEBUG nova.virt.hardware [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 613.272232] env[61868]: DEBUG nova.virt.hardware [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 613.272394] env[61868]: DEBUG nova.virt.hardware [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 613.272575] env[61868]: DEBUG nova.virt.hardware [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 613.272719] env[61868]: DEBUG nova.virt.hardware [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 613.273016] env[61868]: DEBUG nova.virt.hardware [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 613.273271] env[61868]: DEBUG nova.virt.hardware [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 613.273507] env[61868]: DEBUG nova.virt.hardware [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 613.273735] env[61868]: DEBUG nova.virt.hardware [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 613.274082] env[61868]: DEBUG nova.virt.hardware [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 613.274307] env[61868]: DEBUG nova.virt.hardware [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 613.275380] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-599b2a9a-c3dd-4fd3-bdf0-2ef2d6be641e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.285586] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68d1fcb2-33f2-42ea-8719-a7ff3fc43c59 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.301448] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Instance VIF info [] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 613.307020] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Creating folder: Project (61aac352634542dab910ec07b67443b8). Parent ref: group-v281478. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 613.307626] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-09b75044-dfac-4cce-8eca-2689c809efdf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.315896] env[61868]: DEBUG nova.network.neutron [-] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.320336] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Created folder: Project (61aac352634542dab910ec07b67443b8) in parent group-v281478. [ 613.320472] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Creating folder: Instances. Parent ref: group-v281486. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 613.320739] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d5c73ede-5bcc-4bce-a52f-19fa2336da93 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.330994] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Created folder: Instances in parent group-v281486. [ 613.331315] env[61868]: DEBUG oslo.service.loopingcall [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 613.331575] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 613.331725] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-89ada950-3d9e-4604-bf88-a11e74e5232f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.346883] env[61868]: DEBUG oslo_concurrency.lockutils [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.131s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.352032] env[61868]: DEBUG oslo_concurrency.lockutils [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.959s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.352032] env[61868]: INFO nova.compute.claims [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 613.360216] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 613.360216] env[61868]: value = "task-1315277" [ 613.360216] env[61868]: _type = "Task" [ 613.360216] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.369645] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315277, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.384124] env[61868]: INFO nova.scheduler.client.report [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Deleted allocations for instance ddde9ee7-5198-42e5-86b9-f51727dfbf60 [ 613.793582] env[61868]: DEBUG nova.network.neutron [req-045d4c8f-b353-4c0a-a828-0e21af46a6c9 req-6576c2f0-3ca9-4699-9f62-a78af2a1f644 service nova] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.818469] env[61868]: INFO nova.compute.manager [-] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Took 1.04 seconds to deallocate network for instance. [ 613.820823] env[61868]: DEBUG nova.compute.claims [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 613.821114] env[61868]: DEBUG oslo_concurrency.lockutils [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.878017] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315277, 'name': CreateVM_Task, 'duration_secs': 0.329247} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.878017] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 613.878017] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 613.878017] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.878017] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 613.878017] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d193c09-08a6-4940-be38-c9d07c63f6e1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.882774] env[61868]: DEBUG oslo_vmware.api [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Waiting for the task: (returnval){ [ 613.882774] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5267d1b1-cab8-c82a-51f7-41bf4a519984" [ 613.882774] env[61868]: _type = "Task" [ 613.882774] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.897015] env[61868]: DEBUG oslo_vmware.api [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5267d1b1-cab8-c82a-51f7-41bf4a519984, 'name': SearchDatastore_Task, 'duration_secs': 0.009211} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.897520] env[61868]: DEBUG oslo_concurrency.lockutils [None req-810c72ee-4359-45bf-b23a-ba0fc0c7a656 tempest-ServerDiagnosticsV248Test-1102305648 tempest-ServerDiagnosticsV248Test-1102305648-project-member] Lock "ddde9ee7-5198-42e5-86b9-f51727dfbf60" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.949s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.905042] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.905042] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 613.905042] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 613.905042] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.905220] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 613.905220] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a8714da6-9124-4c69-b42d-c8f1e4ec5390 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.911226] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 613.911725] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 613.913263] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e7d2710-cf48-4fd4-a0f1-ec821f5ea463 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.923088] env[61868]: DEBUG oslo_vmware.api [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Waiting for the task: (returnval){ [ 613.923088] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]521bc31c-d618-b604-8c53-beb0a29b8e68" [ 613.923088] env[61868]: _type = "Task" [ 613.923088] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.931940] env[61868]: DEBUG oslo_vmware.api [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]521bc31c-d618-b604-8c53-beb0a29b8e68, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.949099] env[61868]: DEBUG nova.network.neutron [req-045d4c8f-b353-4c0a-a828-0e21af46a6c9 req-6576c2f0-3ca9-4699-9f62-a78af2a1f644 service nova] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.436259] env[61868]: DEBUG oslo_vmware.api [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]521bc31c-d618-b604-8c53-beb0a29b8e68, 'name': SearchDatastore_Task, 'duration_secs': 0.008482} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.437056] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3890c678-a096-4eeb-9362-c35540bd3175 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.445755] env[61868]: DEBUG oslo_vmware.api [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Waiting for the task: (returnval){ [ 614.445755] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52c12081-6ecb-946b-ac89-eba888809585" [ 614.445755] env[61868]: _type = "Task" [ 614.445755] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.455448] env[61868]: DEBUG oslo_concurrency.lockutils [req-045d4c8f-b353-4c0a-a828-0e21af46a6c9 req-6576c2f0-3ca9-4699-9f62-a78af2a1f644 service nova] Releasing lock "refresh_cache-2f9deec0-e0ad-4cb4-9868-9aa53cdf602d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 614.455646] env[61868]: DEBUG nova.compute.manager [req-045d4c8f-b353-4c0a-a828-0e21af46a6c9 req-6576c2f0-3ca9-4699-9f62-a78af2a1f644 service nova] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Received event network-vif-deleted-8354b128-a713-44f1-9205-d5dc691119b0 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 614.455976] env[61868]: DEBUG oslo_vmware.api [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52c12081-6ecb-946b-ac89-eba888809585, 'name': SearchDatastore_Task, 'duration_secs': 0.009318} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.459014] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 614.459014] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 7651f6d2-1825-46a1-9380-732b50abd3cc/7651f6d2-1825-46a1-9380-732b50abd3cc.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 614.459263] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fd5e7f11-0507-4ed9-bac8-db63c9115978 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.467125] env[61868]: DEBUG oslo_vmware.api [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Waiting for the task: (returnval){ [ 614.467125] env[61868]: value = "task-1315278" [ 614.467125] env[61868]: _type = "Task" [ 614.467125] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.479818] env[61868]: DEBUG oslo_vmware.api [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315278, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.822228] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f56ff069-1f6b-4bd1-b66b-f7bde34adb6d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.831231] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15a7758a-3c2c-4d07-a620-6c948ef7c8de {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.873349] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2a44b6f-6ec3-4b44-af71-246c577d3429 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.886904] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccd1ed81-0d10-495c-a35c-d1d74816f3c5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.905602] env[61868]: DEBUG nova.compute.provider_tree [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 614.982721] env[61868]: DEBUG oslo_vmware.api [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315278, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.470141} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.983497] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 7651f6d2-1825-46a1-9380-732b50abd3cc/7651f6d2-1825-46a1-9380-732b50abd3cc.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 614.983497] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 614.983497] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-00dfdc6a-4814-49a4-99bf-d9c8f9fc8111 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.993864] env[61868]: DEBUG oslo_vmware.api [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Waiting for the task: (returnval){ [ 614.993864] env[61868]: value = "task-1315279" [ 614.993864] env[61868]: _type = "Task" [ 614.993864] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.003305] env[61868]: DEBUG oslo_vmware.api [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315279, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.409305] env[61868]: DEBUG nova.scheduler.client.report [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 615.504536] env[61868]: DEBUG oslo_vmware.api [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315279, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074076} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.504857] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 615.505958] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24b53700-76dd-4f99-a781-d12b3401039c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.527053] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Reconfiguring VM instance instance-00000011 to attach disk [datastore1] 7651f6d2-1825-46a1-9380-732b50abd3cc/7651f6d2-1825-46a1-9380-732b50abd3cc.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 615.527348] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d9b44f1c-3509-4686-8676-18ced47a0fdf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.549326] env[61868]: DEBUG oslo_vmware.api [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Waiting for the task: (returnval){ [ 615.549326] env[61868]: value = "task-1315280" [ 615.549326] env[61868]: _type = "Task" [ 615.549326] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.558943] env[61868]: DEBUG oslo_vmware.api [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315280, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.914454] env[61868]: DEBUG oslo_concurrency.lockutils [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.565s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 615.915014] env[61868]: DEBUG nova.compute.manager [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 615.918320] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.060s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.061512] env[61868]: DEBUG oslo_vmware.api [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315280, 'name': ReconfigVM_Task, 'duration_secs': 0.291648} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.061512] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Reconfigured VM instance instance-00000011 to attach disk [datastore1] 7651f6d2-1825-46a1-9380-732b50abd3cc/7651f6d2-1825-46a1-9380-732b50abd3cc.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 616.061838] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d3ba78de-474e-4655-9029-3f28013c9e5f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.069166] env[61868]: DEBUG oslo_vmware.api [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Waiting for the task: (returnval){ [ 616.069166] env[61868]: value = "task-1315281" [ 616.069166] env[61868]: _type = "Task" [ 616.069166] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.080109] env[61868]: DEBUG oslo_vmware.api [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315281, 'name': Rename_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.429925] env[61868]: DEBUG nova.compute.utils [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 616.438377] env[61868]: DEBUG nova.compute.manager [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 616.438560] env[61868]: DEBUG nova.network.neutron [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 616.536930] env[61868]: DEBUG nova.policy [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '330bb66f67894b2ba2b5c33dcb85c9e6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '53ccad64a1ee4f0b8ea982d9d651d8c3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 616.582951] env[61868]: DEBUG oslo_vmware.api [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315281, 'name': Rename_Task, 'duration_secs': 0.143993} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.586019] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 616.586504] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c3818a01-2249-4070-b266-44f0ec1297b1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.596361] env[61868]: DEBUG oslo_vmware.api [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Waiting for the task: (returnval){ [ 616.596361] env[61868]: value = "task-1315282" [ 616.596361] env[61868]: _type = "Task" [ 616.596361] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.627511] env[61868]: DEBUG oslo_vmware.api [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315282, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.941546] env[61868]: DEBUG nova.compute.manager [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 616.954784] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83adba83-9836-4bb0-8bc8-962e2927d78c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.963509] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49d4b696-844e-47b9-b57c-7e75025bbcca {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.997385] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef07d5f4-74e2-40c7-8484-13ed209450e3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.008092] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5d69f5e-d07a-46c5-b01a-68354868817b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.021264] env[61868]: DEBUG nova.compute.provider_tree [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Updating inventory in ProviderTree for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 617.107090] env[61868]: DEBUG oslo_vmware.api [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315282, 'name': PowerOnVM_Task, 'duration_secs': 0.475412} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.107267] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 617.107456] env[61868]: INFO nova.compute.manager [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Took 3.87 seconds to spawn the instance on the hypervisor. [ 617.107626] env[61868]: DEBUG nova.compute.manager [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 617.108396] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef0a25f8-cd72-4813-a3c3-16a481c5cb7d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.366172] env[61868]: DEBUG nova.network.neutron [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Successfully created port: a0af3a45-2562-432e-9a98-d8ed20d76195 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 617.548438] env[61868]: ERROR nova.scheduler.client.report [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] [req-38e14162-f6ae-41aa-ae37-a58d905b0df6] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 83e67721-2ac9-4a23-aa31-82aca86979c8. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-38e14162-f6ae-41aa-ae37-a58d905b0df6"}]}: nova.exception.PortBindingFailed: Binding failed for port d78df532-d2e6-447c-b5d8-9eecf145522c, please check neutron logs for more information. [ 617.569953] env[61868]: DEBUG nova.scheduler.client.report [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Refreshing inventories for resource provider 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 617.589094] env[61868]: DEBUG nova.scheduler.client.report [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Updating ProviderTree inventory for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 617.589094] env[61868]: DEBUG nova.compute.provider_tree [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Updating inventory in ProviderTree for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 617.607776] env[61868]: DEBUG nova.scheduler.client.report [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Refreshing aggregate associations for resource provider 83e67721-2ac9-4a23-aa31-82aca86979c8, aggregates: None {{(pid=61868) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 617.626863] env[61868]: INFO nova.compute.manager [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Took 26.11 seconds to build instance. [ 617.631481] env[61868]: DEBUG nova.scheduler.client.report [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Refreshing trait associations for resource provider 83e67721-2ac9-4a23-aa31-82aca86979c8, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61868) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 617.952017] env[61868]: DEBUG nova.compute.manager [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 617.991890] env[61868]: DEBUG nova.virt.hardware [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 617.992224] env[61868]: DEBUG nova.virt.hardware [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 617.992381] env[61868]: DEBUG nova.virt.hardware [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 617.992559] env[61868]: DEBUG nova.virt.hardware [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 617.992769] env[61868]: DEBUG nova.virt.hardware [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 617.992845] env[61868]: DEBUG nova.virt.hardware [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 617.993276] env[61868]: DEBUG nova.virt.hardware [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 617.993429] env[61868]: DEBUG nova.virt.hardware [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 617.993600] env[61868]: DEBUG nova.virt.hardware [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 617.993759] env[61868]: DEBUG nova.virt.hardware [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 617.993922] env[61868]: DEBUG nova.virt.hardware [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 617.995099] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be00090f-beb0-4044-a09e-41033b81ca96 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.009848] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9448b3d2-9c7c-475d-9507-f0c3651a9022 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.106262] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47268501-50cc-4996-9739-72db89cd9c6e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.116689] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-262b9134-0ee4-49ea-b3da-fb6b61679ecb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.167369] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bd71e4ef-f41f-4f4a-b23f-91f5b5a3494a tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Lock "7651f6d2-1825-46a1-9380-732b50abd3cc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 56.132s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.169610] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93b767b8-6a76-446e-a97c-d9f48c5ffa2f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.180192] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb098e76-c30f-41a9-9b22-71506e7403f0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.200019] env[61868]: DEBUG nova.compute.provider_tree [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Updating inventory in ProviderTree for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 618.337830] env[61868]: DEBUG nova.compute.manager [req-a43d1ee5-3bef-4fd9-83ff-c82fa72833b9 req-a719e7c8-13af-43ac-a86a-80e987597c3c service nova] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Received event network-changed-a0af3a45-2562-432e-9a98-d8ed20d76195 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 618.338708] env[61868]: DEBUG nova.compute.manager [req-a43d1ee5-3bef-4fd9-83ff-c82fa72833b9 req-a719e7c8-13af-43ac-a86a-80e987597c3c service nova] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Refreshing instance network info cache due to event network-changed-a0af3a45-2562-432e-9a98-d8ed20d76195. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 618.338876] env[61868]: DEBUG oslo_concurrency.lockutils [req-a43d1ee5-3bef-4fd9-83ff-c82fa72833b9 req-a719e7c8-13af-43ac-a86a-80e987597c3c service nova] Acquiring lock "refresh_cache-8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.339041] env[61868]: DEBUG oslo_concurrency.lockutils [req-a43d1ee5-3bef-4fd9-83ff-c82fa72833b9 req-a719e7c8-13af-43ac-a86a-80e987597c3c service nova] Acquired lock "refresh_cache-8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.339209] env[61868]: DEBUG nova.network.neutron [req-a43d1ee5-3bef-4fd9-83ff-c82fa72833b9 req-a719e7c8-13af-43ac-a86a-80e987597c3c service nova] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Refreshing network info cache for port a0af3a45-2562-432e-9a98-d8ed20d76195 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 618.483087] env[61868]: ERROR nova.compute.manager [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a0af3a45-2562-432e-9a98-d8ed20d76195, please check neutron logs for more information. [ 618.483087] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 618.483087] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.483087] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 618.483087] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 618.483087] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 618.483087] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 618.483087] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 618.483087] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.483087] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 618.483087] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.483087] env[61868]: ERROR nova.compute.manager raise self.value [ 618.483087] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 618.483087] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 618.483087] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.483087] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 618.483485] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.483485] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 618.483485] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a0af3a45-2562-432e-9a98-d8ed20d76195, please check neutron logs for more information. [ 618.483485] env[61868]: ERROR nova.compute.manager [ 618.483485] env[61868]: Traceback (most recent call last): [ 618.483485] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 618.483485] env[61868]: listener.cb(fileno) [ 618.483485] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.483485] env[61868]: result = function(*args, **kwargs) [ 618.483485] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 618.483485] env[61868]: return func(*args, **kwargs) [ 618.483485] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 618.483485] env[61868]: raise e [ 618.483485] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.483485] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 618.483485] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 618.483485] env[61868]: created_port_ids = self._update_ports_for_instance( [ 618.483485] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 618.483485] env[61868]: with excutils.save_and_reraise_exception(): [ 618.483485] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.483485] env[61868]: self.force_reraise() [ 618.483485] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.483485] env[61868]: raise self.value [ 618.483485] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 618.483485] env[61868]: updated_port = self._update_port( [ 618.483485] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.483485] env[61868]: _ensure_no_port_binding_failure(port) [ 618.483485] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.483485] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 618.484121] env[61868]: nova.exception.PortBindingFailed: Binding failed for port a0af3a45-2562-432e-9a98-d8ed20d76195, please check neutron logs for more information. [ 618.484121] env[61868]: Removing descriptor: 15 [ 618.484121] env[61868]: ERROR nova.compute.manager [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a0af3a45-2562-432e-9a98-d8ed20d76195, please check neutron logs for more information. [ 618.484121] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Traceback (most recent call last): [ 618.484121] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 618.484121] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] yield resources [ 618.484121] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 618.484121] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] self.driver.spawn(context, instance, image_meta, [ 618.484121] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 618.484121] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] self._vmops.spawn(context, instance, image_meta, injected_files, [ 618.484121] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 618.484121] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] vm_ref = self.build_virtual_machine(instance, [ 618.484382] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 618.484382] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] vif_infos = vmwarevif.get_vif_info(self._session, [ 618.484382] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 618.484382] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] for vif in network_info: [ 618.484382] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 618.484382] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] return self._sync_wrapper(fn, *args, **kwargs) [ 618.484382] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 618.484382] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] self.wait() [ 618.484382] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 618.484382] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] self[:] = self._gt.wait() [ 618.484382] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 618.484382] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] return self._exit_event.wait() [ 618.484382] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 618.484720] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] result = hub.switch() [ 618.484720] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 618.484720] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] return self.greenlet.switch() [ 618.484720] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.484720] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] result = function(*args, **kwargs) [ 618.484720] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 618.484720] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] return func(*args, **kwargs) [ 618.484720] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 618.484720] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] raise e [ 618.484720] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.484720] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] nwinfo = self.network_api.allocate_for_instance( [ 618.484720] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 618.484720] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] created_port_ids = self._update_ports_for_instance( [ 618.484997] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 618.484997] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] with excutils.save_and_reraise_exception(): [ 618.484997] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.484997] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] self.force_reraise() [ 618.484997] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.484997] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] raise self.value [ 618.484997] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 618.484997] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] updated_port = self._update_port( [ 618.484997] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.484997] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] _ensure_no_port_binding_failure(port) [ 618.484997] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.484997] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] raise exception.PortBindingFailed(port_id=port['id']) [ 618.485262] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] nova.exception.PortBindingFailed: Binding failed for port a0af3a45-2562-432e-9a98-d8ed20d76195, please check neutron logs for more information. [ 618.485262] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] [ 618.485262] env[61868]: INFO nova.compute.manager [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Terminating instance [ 618.490727] env[61868]: DEBUG oslo_concurrency.lockutils [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Acquiring lock "refresh_cache-8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.673682] env[61868]: DEBUG nova.compute.manager [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 618.749122] env[61868]: DEBUG nova.scheduler.client.report [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Updated inventory for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 with generation 44 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 618.749122] env[61868]: DEBUG nova.compute.provider_tree [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Updating resource provider 83e67721-2ac9-4a23-aa31-82aca86979c8 generation from 44 to 45 during operation: update_inventory {{(pid=61868) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 618.749122] env[61868]: DEBUG nova.compute.provider_tree [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Updating inventory in ProviderTree for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 618.868702] env[61868]: DEBUG nova.network.neutron [req-a43d1ee5-3bef-4fd9-83ff-c82fa72833b9 req-a719e7c8-13af-43ac-a86a-80e987597c3c service nova] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 618.923810] env[61868]: DEBUG nova.network.neutron [req-a43d1ee5-3bef-4fd9-83ff-c82fa72833b9 req-a719e7c8-13af-43ac-a86a-80e987597c3c service nova] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.054897] env[61868]: INFO nova.compute.manager [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Rebuilding instance [ 619.110716] env[61868]: DEBUG nova.compute.manager [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 619.112622] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23ffac12-1d1e-4e7d-8e0e-763c58e83d5f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.195178] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.254787] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 3.335s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.257147] env[61868]: ERROR nova.compute.manager [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d78df532-d2e6-447c-b5d8-9eecf145522c, please check neutron logs for more information. [ 619.257147] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Traceback (most recent call last): [ 619.257147] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 619.257147] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] self.driver.spawn(context, instance, image_meta, [ 619.257147] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 619.257147] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] self._vmops.spawn(context, instance, image_meta, injected_files, [ 619.257147] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 619.257147] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] vm_ref = self.build_virtual_machine(instance, [ 619.257147] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 619.257147] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] vif_infos = vmwarevif.get_vif_info(self._session, [ 619.257147] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 619.257606] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] for vif in network_info: [ 619.257606] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 619.257606] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] return self._sync_wrapper(fn, *args, **kwargs) [ 619.257606] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 619.257606] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] self.wait() [ 619.257606] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 619.257606] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] self[:] = self._gt.wait() [ 619.257606] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 619.257606] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] return self._exit_event.wait() [ 619.257606] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 619.257606] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] result = hub.switch() [ 619.257606] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 619.257606] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] return self.greenlet.switch() [ 619.258248] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 619.258248] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] result = function(*args, **kwargs) [ 619.258248] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 619.258248] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] return func(*args, **kwargs) [ 619.258248] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 619.258248] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] raise e [ 619.258248] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 619.258248] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] nwinfo = self.network_api.allocate_for_instance( [ 619.258248] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 619.258248] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] created_port_ids = self._update_ports_for_instance( [ 619.258248] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 619.258248] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] with excutils.save_and_reraise_exception(): [ 619.258248] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 619.258663] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] self.force_reraise() [ 619.258663] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 619.258663] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] raise self.value [ 619.258663] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 619.258663] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] updated_port = self._update_port( [ 619.258663] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 619.258663] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] _ensure_no_port_binding_failure(port) [ 619.258663] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 619.258663] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] raise exception.PortBindingFailed(port_id=port['id']) [ 619.258663] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] nova.exception.PortBindingFailed: Binding failed for port d78df532-d2e6-447c-b5d8-9eecf145522c, please check neutron logs for more information. [ 619.258663] env[61868]: ERROR nova.compute.manager [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] [ 619.259040] env[61868]: DEBUG nova.compute.utils [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Binding failed for port d78df532-d2e6-447c-b5d8-9eecf145522c, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 619.259040] env[61868]: DEBUG oslo_concurrency.lockutils [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.013s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.260556] env[61868]: INFO nova.compute.claims [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 619.265061] env[61868]: DEBUG nova.compute.manager [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Build of instance a499452b-9d48-4022-8d4c-0c66c6dde9ba was re-scheduled: Binding failed for port d78df532-d2e6-447c-b5d8-9eecf145522c, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 619.265781] env[61868]: DEBUG nova.compute.manager [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 619.266125] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Acquiring lock "refresh_cache-a499452b-9d48-4022-8d4c-0c66c6dde9ba" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.266270] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Acquired lock "refresh_cache-a499452b-9d48-4022-8d4c-0c66c6dde9ba" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.266439] env[61868]: DEBUG nova.network.neutron [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 619.429871] env[61868]: DEBUG oslo_concurrency.lockutils [req-a43d1ee5-3bef-4fd9-83ff-c82fa72833b9 req-a719e7c8-13af-43ac-a86a-80e987597c3c service nova] Releasing lock "refresh_cache-8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.430317] env[61868]: DEBUG oslo_concurrency.lockutils [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Acquired lock "refresh_cache-8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.430540] env[61868]: DEBUG nova.network.neutron [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 619.629066] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 619.629066] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f047e3ba-a00f-4996-aeca-f917a628c9e6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.636122] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Waiting for the task: (returnval){ [ 619.636122] env[61868]: value = "task-1315283" [ 619.636122] env[61868]: _type = "Task" [ 619.636122] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.646955] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315283, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.800011] env[61868]: DEBUG nova.network.neutron [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 619.904253] env[61868]: DEBUG nova.network.neutron [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.967240] env[61868]: DEBUG nova.network.neutron [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.112524] env[61868]: DEBUG nova.network.neutron [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.151364] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315283, 'name': PowerOffVM_Task, 'duration_secs': 0.234679} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.151364] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 620.151364] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 620.151364] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7de47a44-9f80-42fb-9cb2-a1d7e5396e84 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.159719] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 620.159719] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e126bfdd-cfdb-4b40-b060-969fa82c6831 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.189371] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 620.189608] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 620.189819] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Deleting the datastore file [datastore1] 7651f6d2-1825-46a1-9380-732b50abd3cc {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 620.190220] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1a7a05c5-f67e-4edb-ae20-dc244c209eb7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.201421] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Waiting for the task: (returnval){ [ 620.201421] env[61868]: value = "task-1315285" [ 620.201421] env[61868]: _type = "Task" [ 620.201421] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.212186] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315285, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.406178] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Releasing lock "refresh_cache-a499452b-9d48-4022-8d4c-0c66c6dde9ba" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.406380] env[61868]: DEBUG nova.compute.manager [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 620.407369] env[61868]: DEBUG nova.compute.manager [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 620.407632] env[61868]: DEBUG nova.network.neutron [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 620.440552] env[61868]: DEBUG nova.network.neutron [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.458194] env[61868]: DEBUG nova.compute.manager [req-9e85f14d-48f4-4dc1-b854-539d54a21321 req-50be224c-59f8-4caa-af1f-af29aca74408 service nova] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Received event network-vif-deleted-a0af3a45-2562-432e-9a98-d8ed20d76195 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 620.615928] env[61868]: DEBUG oslo_concurrency.lockutils [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Releasing lock "refresh_cache-8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.616370] env[61868]: DEBUG nova.compute.manager [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 620.616556] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 620.616853] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4e4948b7-891e-4536-8c07-2b7d657266e5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.633666] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eb5b678-412f-4bba-baca-16fcd11bc84c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.668027] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431 could not be found. [ 620.668027] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 620.668027] env[61868]: INFO nova.compute.manager [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Took 0.05 seconds to destroy the instance on the hypervisor. [ 620.668027] env[61868]: DEBUG oslo.service.loopingcall [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 620.670012] env[61868]: DEBUG nova.compute.manager [-] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 620.670086] env[61868]: DEBUG nova.network.neutron [-] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 620.686574] env[61868]: DEBUG nova.network.neutron [-] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.716573] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315285, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.118323} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.716841] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 620.717032] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 620.717220] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 620.826414] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdcc70b8-ca91-468b-9cc8-f71b0c81720e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.835181] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d85c92b4-e867-448d-8bd2-cf0a8a28a395 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.867654] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab553606-8f6e-4bb8-94af-36474a5b67db {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.878245] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02e4bddd-5829-4e64-b276-7972e291122a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.891443] env[61868]: DEBUG nova.compute.provider_tree [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 620.945668] env[61868]: DEBUG nova.network.neutron [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.190170] env[61868]: DEBUG nova.network.neutron [-] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.396170] env[61868]: DEBUG nova.scheduler.client.report [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 621.453516] env[61868]: INFO nova.compute.manager [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] [instance: a499452b-9d48-4022-8d4c-0c66c6dde9ba] Took 1.04 seconds to deallocate network for instance. [ 621.694869] env[61868]: INFO nova.compute.manager [-] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Took 1.02 seconds to deallocate network for instance. [ 621.697050] env[61868]: DEBUG nova.compute.claims [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 621.697050] env[61868]: DEBUG oslo_concurrency.lockutils [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.757068] env[61868]: DEBUG nova.virt.hardware [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 621.757068] env[61868]: DEBUG nova.virt.hardware [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 621.757068] env[61868]: DEBUG nova.virt.hardware [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 621.760109] env[61868]: DEBUG nova.virt.hardware [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 621.760109] env[61868]: DEBUG nova.virt.hardware [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 621.760109] env[61868]: DEBUG nova.virt.hardware [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 621.760109] env[61868]: DEBUG nova.virt.hardware [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 621.760109] env[61868]: DEBUG nova.virt.hardware [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 621.760603] env[61868]: DEBUG nova.virt.hardware [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 621.760603] env[61868]: DEBUG nova.virt.hardware [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 621.760603] env[61868]: DEBUG nova.virt.hardware [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 621.760603] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdcdee08-759f-413c-92fd-cab99f267609 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.769143] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-583f884b-7b1c-4fef-9ae8-2d3bd565a29a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.782427] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Instance VIF info [] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 621.788083] env[61868]: DEBUG oslo.service.loopingcall [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 621.788333] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 621.788542] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-22a2dfc9-cdf9-45b7-a5df-348b171ce939 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.807515] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 621.807515] env[61868]: value = "task-1315286" [ 621.807515] env[61868]: _type = "Task" [ 621.807515] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.817619] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315286, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.900946] env[61868]: DEBUG oslo_concurrency.lockutils [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.643s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.901637] env[61868]: DEBUG nova.compute.manager [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 621.904962] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.396s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.133910] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] Acquiring lock "97d1d9c9-214f-4573-9efa-3254ea3f2ee0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.134160] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] Lock "97d1d9c9-214f-4573-9efa-3254ea3f2ee0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.321504] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315286, 'name': CreateVM_Task, 'duration_secs': 0.292484} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.321681] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 622.322697] env[61868]: DEBUG oslo_vmware.service [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b7ffcfd-43ff-4954-8cf0-9274f0ebbf16 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.329603] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 622.330704] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.331096] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 622.331357] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4887f4d-a48b-4fa2-90ca-9dfb6aa8c121 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.336544] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Waiting for the task: (returnval){ [ 622.336544] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5216108f-f863-88ac-6cea-9108258738d0" [ 622.336544] env[61868]: _type = "Task" [ 622.336544] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.344608] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5216108f-f863-88ac-6cea-9108258738d0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.412718] env[61868]: DEBUG nova.compute.utils [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 622.414075] env[61868]: DEBUG nova.compute.manager [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 622.414247] env[61868]: DEBUG nova.network.neutron [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 622.488133] env[61868]: INFO nova.scheduler.client.report [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Deleted allocations for instance a499452b-9d48-4022-8d4c-0c66c6dde9ba [ 622.503100] env[61868]: DEBUG nova.policy [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '330bb66f67894b2ba2b5c33dcb85c9e6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '53ccad64a1ee4f0b8ea982d9d651d8c3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 622.862531] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 622.862531] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 622.862531] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 622.862531] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.862659] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 622.862659] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b472d0fa-9fa6-4da7-aa88-c280bcdde0b8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.871772] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 622.872114] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 622.874107] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcf6e14a-2c5d-4de0-90e1-e7faa45eb1f3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.883280] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b7941a85-d0bc-4848-be94-f79b8d00f57d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.889925] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb6293c9-0106-48bb-ba78-a6bf0eade695 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.896157] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07e8cb3c-9cc9-4087-abeb-0cb5886f9a4e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.899535] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Waiting for the task: (returnval){ [ 622.899535] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]523d1ee0-817e-865a-6788-80beede3f10a" [ 622.899535] env[61868]: _type = "Task" [ 622.899535] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.937431] env[61868]: DEBUG nova.compute.manager [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 622.944284] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d362a294-674e-4f3a-9d52-e79edf9aecf5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.946597] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]523d1ee0-817e-865a-6788-80beede3f10a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.954907] env[61868]: DEBUG nova.network.neutron [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Successfully created port: d9e16fa5-610c-4053-8a20-1c33bf8d6324 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 622.957946] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c455274-fc30-4de7-ac7d-8f23da23da42 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.975054] env[61868]: DEBUG nova.compute.provider_tree [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 622.998764] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f88e82fe-fa82-476c-9a46-9a23d3ec3c6d tempest-InstanceActionsTestJSON-614902237 tempest-InstanceActionsTestJSON-614902237-project-member] Lock "a499452b-9d48-4022-8d4c-0c66c6dde9ba" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.763s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.187105] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Acquiring lock "84dc8092-c24a-4587-bfcb-865b6eb129aa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.187343] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Lock "84dc8092-c24a-4587-bfcb-865b6eb129aa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.412830] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Preparing fetch location {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 623.413110] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Creating directory with path [datastore2] vmware_temp/89d7d0cc-eef1-42bf-97df-1d0615b8fcd1/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 623.413416] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-db971d78-af18-4e41-99e6-c1758d7d75f8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.430593] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Created directory with path [datastore2] vmware_temp/89d7d0cc-eef1-42bf-97df-1d0615b8fcd1/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 623.430806] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Fetch image to [datastore2] vmware_temp/89d7d0cc-eef1-42bf-97df-1d0615b8fcd1/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/tmp-sparse.vmdk {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 623.430974] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Downloading image file data c4fdfa42-c9a0-41ca-801e-4e55c3d4328d to [datastore2] vmware_temp/89d7d0cc-eef1-42bf-97df-1d0615b8fcd1/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/tmp-sparse.vmdk on the data store datastore2 {{(pid=61868) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 623.431746] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a11f815c-2033-46e0-bd1d-e0b58e89c767 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.440196] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba17bbea-ef68-4ec0-96db-94a86741fa82 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.459559] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-666322d8-5a69-4639-af75-3fcf507c59e3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.500553] env[61868]: DEBUG nova.scheduler.client.report [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 623.506789] env[61868]: DEBUG nova.compute.manager [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 623.511238] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba5d071a-d4ae-4ea7-ac55-28f287a86d14 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.518871] env[61868]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-333b4326-8ff0-479f-a473-9e742905bd1a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.557719] env[61868]: DEBUG nova.virt.vmwareapi.images [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Downloading image file data c4fdfa42-c9a0-41ca-801e-4e55c3d4328d to the data store datastore2 {{(pid=61868) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 623.597043] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Acquiring lock "06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.597450] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Lock "06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.623573] env[61868]: DEBUG oslo_vmware.rw_handles [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/89d7d0cc-eef1-42bf-97df-1d0615b8fcd1/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61868) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 623.965623] env[61868]: DEBUG nova.compute.manager [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 623.997826] env[61868]: DEBUG nova.virt.hardware [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 623.998091] env[61868]: DEBUG nova.virt.hardware [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 623.998246] env[61868]: DEBUG nova.virt.hardware [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 623.998413] env[61868]: DEBUG nova.virt.hardware [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 623.998594] env[61868]: DEBUG nova.virt.hardware [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 623.998745] env[61868]: DEBUG nova.virt.hardware [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 623.998985] env[61868]: DEBUG nova.virt.hardware [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 623.999327] env[61868]: DEBUG nova.virt.hardware [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 623.999523] env[61868]: DEBUG nova.virt.hardware [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 623.999928] env[61868]: DEBUG nova.virt.hardware [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 624.000180] env[61868]: DEBUG nova.virt.hardware [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 624.001576] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f53bf902-dcfd-4461-84ff-19e6d1dbf18c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.009949] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.104s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 624.009949] env[61868]: ERROR nova.compute.manager [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e00fd5a7-bf99-4fb6-a8c4-11547f313ca3, please check neutron logs for more information. [ 624.009949] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] Traceback (most recent call last): [ 624.009949] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 624.009949] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] self.driver.spawn(context, instance, image_meta, [ 624.009949] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 624.009949] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] self._vmops.spawn(context, instance, image_meta, injected_files, [ 624.009949] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 624.009949] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] vm_ref = self.build_virtual_machine(instance, [ 624.010266] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 624.010266] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] vif_infos = vmwarevif.get_vif_info(self._session, [ 624.010266] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 624.010266] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] for vif in network_info: [ 624.010266] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 624.010266] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] return self._sync_wrapper(fn, *args, **kwargs) [ 624.010266] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 624.010266] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] self.wait() [ 624.010266] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 624.010266] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] self[:] = self._gt.wait() [ 624.010266] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 624.010266] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] return self._exit_event.wait() [ 624.010266] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 624.010540] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] current.throw(*self._exc) [ 624.010540] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 624.010540] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] result = function(*args, **kwargs) [ 624.010540] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 624.010540] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] return func(*args, **kwargs) [ 624.010540] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 624.010540] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] raise e [ 624.010540] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.010540] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] nwinfo = self.network_api.allocate_for_instance( [ 624.010540] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 624.010540] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] created_port_ids = self._update_ports_for_instance( [ 624.010540] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 624.010540] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] with excutils.save_and_reraise_exception(): [ 624.010814] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.010814] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] self.force_reraise() [ 624.010814] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.010814] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] raise self.value [ 624.010814] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 624.010814] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] updated_port = self._update_port( [ 624.010814] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.010814] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] _ensure_no_port_binding_failure(port) [ 624.010814] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.010814] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] raise exception.PortBindingFailed(port_id=port['id']) [ 624.010814] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] nova.exception.PortBindingFailed: Binding failed for port e00fd5a7-bf99-4fb6-a8c4-11547f313ca3, please check neutron logs for more information. [ 624.010814] env[61868]: ERROR nova.compute.manager [instance: 5e812408-6f27-4b52-9564-4063f4526340] [ 624.011731] env[61868]: DEBUG nova.compute.utils [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Binding failed for port e00fd5a7-bf99-4fb6-a8c4-11547f313ca3, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 624.013461] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.014s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.015833] env[61868]: INFO nova.compute.claims [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 624.022657] env[61868]: DEBUG nova.compute.manager [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Build of instance 5e812408-6f27-4b52-9564-4063f4526340 was re-scheduled: Binding failed for port e00fd5a7-bf99-4fb6-a8c4-11547f313ca3, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 624.023212] env[61868]: DEBUG nova.compute.manager [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 624.023451] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] Acquiring lock "refresh_cache-5e812408-6f27-4b52-9564-4063f4526340" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.023511] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] Acquired lock "refresh_cache-5e812408-6f27-4b52-9564-4063f4526340" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.026140] env[61868]: DEBUG nova.network.neutron [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 624.030093] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ef0093a-ce8a-4067-9a7c-73a1f8ab5a5a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.059626] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Acquiring lock "2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.059872] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Lock "2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.079721] env[61868]: ERROR nova.compute.manager [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d9e16fa5-610c-4053-8a20-1c33bf8d6324, please check neutron logs for more information. [ 624.079721] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 624.079721] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.079721] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 624.079721] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 624.079721] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 624.079721] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 624.079721] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 624.079721] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.079721] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 624.079721] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.079721] env[61868]: ERROR nova.compute.manager raise self.value [ 624.079721] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 624.079721] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 624.079721] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.079721] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 624.081288] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.081288] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 624.081288] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d9e16fa5-610c-4053-8a20-1c33bf8d6324, please check neutron logs for more information. [ 624.081288] env[61868]: ERROR nova.compute.manager [ 624.081288] env[61868]: Traceback (most recent call last): [ 624.081288] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 624.081288] env[61868]: listener.cb(fileno) [ 624.081288] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 624.081288] env[61868]: result = function(*args, **kwargs) [ 624.081288] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 624.081288] env[61868]: return func(*args, **kwargs) [ 624.081288] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 624.081288] env[61868]: raise e [ 624.081288] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.081288] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 624.081288] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 624.081288] env[61868]: created_port_ids = self._update_ports_for_instance( [ 624.081288] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 624.081288] env[61868]: with excutils.save_and_reraise_exception(): [ 624.081288] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.081288] env[61868]: self.force_reraise() [ 624.081288] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.081288] env[61868]: raise self.value [ 624.081288] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 624.081288] env[61868]: updated_port = self._update_port( [ 624.081288] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.081288] env[61868]: _ensure_no_port_binding_failure(port) [ 624.081288] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.081288] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 624.081976] env[61868]: nova.exception.PortBindingFailed: Binding failed for port d9e16fa5-610c-4053-8a20-1c33bf8d6324, please check neutron logs for more information. [ 624.081976] env[61868]: Removing descriptor: 15 [ 624.081976] env[61868]: ERROR nova.compute.manager [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d9e16fa5-610c-4053-8a20-1c33bf8d6324, please check neutron logs for more information. [ 624.081976] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Traceback (most recent call last): [ 624.081976] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 624.081976] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] yield resources [ 624.081976] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 624.081976] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] self.driver.spawn(context, instance, image_meta, [ 624.081976] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 624.081976] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 624.081976] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 624.081976] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] vm_ref = self.build_virtual_machine(instance, [ 624.082267] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 624.082267] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] vif_infos = vmwarevif.get_vif_info(self._session, [ 624.082267] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 624.082267] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] for vif in network_info: [ 624.082267] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 624.082267] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] return self._sync_wrapper(fn, *args, **kwargs) [ 624.082267] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 624.082267] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] self.wait() [ 624.082267] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 624.082267] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] self[:] = self._gt.wait() [ 624.082267] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 624.082267] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] return self._exit_event.wait() [ 624.082267] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 624.082538] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] result = hub.switch() [ 624.082538] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 624.082538] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] return self.greenlet.switch() [ 624.082538] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 624.082538] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] result = function(*args, **kwargs) [ 624.082538] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 624.082538] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] return func(*args, **kwargs) [ 624.082538] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 624.082538] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] raise e [ 624.082538] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.082538] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] nwinfo = self.network_api.allocate_for_instance( [ 624.082538] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 624.082538] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] created_port_ids = self._update_ports_for_instance( [ 624.082803] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 624.082803] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] with excutils.save_and_reraise_exception(): [ 624.082803] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.082803] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] self.force_reraise() [ 624.082803] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.082803] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] raise self.value [ 624.082803] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 624.082803] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] updated_port = self._update_port( [ 624.082803] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.082803] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] _ensure_no_port_binding_failure(port) [ 624.082803] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.082803] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] raise exception.PortBindingFailed(port_id=port['id']) [ 624.083082] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] nova.exception.PortBindingFailed: Binding failed for port d9e16fa5-610c-4053-8a20-1c33bf8d6324, please check neutron logs for more information. [ 624.083082] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] [ 624.083082] env[61868]: INFO nova.compute.manager [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Terminating instance [ 624.084269] env[61868]: DEBUG oslo_concurrency.lockutils [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Acquiring lock "refresh_cache-7c2937c6-e6b7-4716-80fa-5538641bc8dc" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.084470] env[61868]: DEBUG oslo_concurrency.lockutils [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Acquired lock "refresh_cache-7c2937c6-e6b7-4716-80fa-5538641bc8dc" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.084608] env[61868]: DEBUG nova.network.neutron [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 624.096176] env[61868]: DEBUG oslo_concurrency.lockutils [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.231545] env[61868]: DEBUG nova.compute.manager [req-afa382e0-b841-4629-ba1a-3e7a5ead2a6f req-bdbf1ea3-10bf-47f2-ac62-7fb97d3d7f09 service nova] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Received event network-changed-d9e16fa5-610c-4053-8a20-1c33bf8d6324 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 624.231777] env[61868]: DEBUG nova.compute.manager [req-afa382e0-b841-4629-ba1a-3e7a5ead2a6f req-bdbf1ea3-10bf-47f2-ac62-7fb97d3d7f09 service nova] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Refreshing instance network info cache due to event network-changed-d9e16fa5-610c-4053-8a20-1c33bf8d6324. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 624.231970] env[61868]: DEBUG oslo_concurrency.lockutils [req-afa382e0-b841-4629-ba1a-3e7a5ead2a6f req-bdbf1ea3-10bf-47f2-ac62-7fb97d3d7f09 service nova] Acquiring lock "refresh_cache-7c2937c6-e6b7-4716-80fa-5538641bc8dc" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.274363] env[61868]: DEBUG oslo_vmware.rw_handles [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Completed reading data from the image iterator. {{(pid=61868) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 624.274363] env[61868]: DEBUG oslo_vmware.rw_handles [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Closing write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/89d7d0cc-eef1-42bf-97df-1d0615b8fcd1/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61868) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 624.416510] env[61868]: DEBUG nova.virt.vmwareapi.images [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Downloaded image file data c4fdfa42-c9a0-41ca-801e-4e55c3d4328d to vmware_temp/89d7d0cc-eef1-42bf-97df-1d0615b8fcd1/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/tmp-sparse.vmdk on the data store datastore2 {{(pid=61868) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 624.418829] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Caching image {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 624.419098] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Copying Virtual Disk [datastore2] vmware_temp/89d7d0cc-eef1-42bf-97df-1d0615b8fcd1/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/tmp-sparse.vmdk to [datastore2] vmware_temp/89d7d0cc-eef1-42bf-97df-1d0615b8fcd1/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 624.419378] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-863a279a-5fe1-45e9-a88f-f40a882343d2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.429230] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Waiting for the task: (returnval){ [ 624.429230] env[61868]: value = "task-1315287" [ 624.429230] env[61868]: _type = "Task" [ 624.429230] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.438105] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315287, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.675076] env[61868]: DEBUG oslo_concurrency.lockutils [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] Acquiring lock "df927c0c-b742-4451-9439-60ea63287b99" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.675320] env[61868]: DEBUG oslo_concurrency.lockutils [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] Lock "df927c0c-b742-4451-9439-60ea63287b99" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.678107] env[61868]: DEBUG nova.network.neutron [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 624.684018] env[61868]: DEBUG nova.network.neutron [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 624.825051] env[61868]: DEBUG nova.network.neutron [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.833517] env[61868]: DEBUG nova.network.neutron [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.940642] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315287, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.337829] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] Releasing lock "refresh_cache-5e812408-6f27-4b52-9564-4063f4526340" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.339019] env[61868]: DEBUG nova.compute.manager [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 625.339319] env[61868]: DEBUG nova.compute.manager [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 625.339897] env[61868]: DEBUG nova.network.neutron [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] [instance: 5e812408-6f27-4b52-9564-4063f4526340] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 625.345159] env[61868]: DEBUG oslo_concurrency.lockutils [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Releasing lock "refresh_cache-7c2937c6-e6b7-4716-80fa-5538641bc8dc" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.345381] env[61868]: DEBUG nova.compute.manager [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 625.346490] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 625.346490] env[61868]: DEBUG oslo_concurrency.lockutils [req-afa382e0-b841-4629-ba1a-3e7a5ead2a6f req-bdbf1ea3-10bf-47f2-ac62-7fb97d3d7f09 service nova] Acquired lock "refresh_cache-7c2937c6-e6b7-4716-80fa-5538641bc8dc" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.346654] env[61868]: DEBUG nova.network.neutron [req-afa382e0-b841-4629-ba1a-3e7a5ead2a6f req-bdbf1ea3-10bf-47f2-ac62-7fb97d3d7f09 service nova] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Refreshing network info cache for port d9e16fa5-610c-4053-8a20-1c33bf8d6324 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 625.347969] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-50e56b34-36bf-4d3f-bb74-d3773858a899 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.365120] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0363794-8267-47a3-bf31-c6fcb234b88e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.384038] env[61868]: DEBUG nova.network.neutron [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.402341] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7c2937c6-e6b7-4716-80fa-5538641bc8dc could not be found. [ 625.402341] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 625.402513] env[61868]: INFO nova.compute.manager [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Took 0.06 seconds to destroy the instance on the hypervisor. [ 625.404587] env[61868]: DEBUG oslo.service.loopingcall [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 625.405693] env[61868]: DEBUG nova.compute.manager [-] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 625.405789] env[61868]: DEBUG nova.network.neutron [-] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 625.439227] env[61868]: DEBUG nova.network.neutron [-] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.450104] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315287, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.676678} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.450104] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Copied Virtual Disk [datastore2] vmware_temp/89d7d0cc-eef1-42bf-97df-1d0615b8fcd1/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/tmp-sparse.vmdk to [datastore2] vmware_temp/89d7d0cc-eef1-42bf-97df-1d0615b8fcd1/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 625.450104] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Deleting the datastore file [datastore2] vmware_temp/89d7d0cc-eef1-42bf-97df-1d0615b8fcd1/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/tmp-sparse.vmdk {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 625.450104] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e6714102-bab2-43de-9151-a956cbda49a9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.457401] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Waiting for the task: (returnval){ [ 625.457401] env[61868]: value = "task-1315288" [ 625.457401] env[61868]: _type = "Task" [ 625.457401] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.474039] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315288, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.558397] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c00d597-ee2a-4437-a963-610523e8fb89 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.568097] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2305d0e-b103-4cb0-a928-62712f843fd3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.599083] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3c88a62-188c-4333-8acb-25f711a9fb88 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.608426] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c987a5d1-bae4-4df2-9c73-7b1a9e67d56e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.623397] env[61868]: DEBUG nova.compute.provider_tree [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 625.887391] env[61868]: DEBUG nova.network.neutron [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.889683] env[61868]: DEBUG nova.network.neutron [req-afa382e0-b841-4629-ba1a-3e7a5ead2a6f req-bdbf1ea3-10bf-47f2-ac62-7fb97d3d7f09 service nova] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.941228] env[61868]: DEBUG nova.network.neutron [-] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.968656] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315288, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.0251} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.968968] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 625.969257] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Moving file from [datastore2] vmware_temp/89d7d0cc-eef1-42bf-97df-1d0615b8fcd1/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d to [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d. {{(pid=61868) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 625.969562] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-74f41b16-868a-412b-96f9-f613c01b0ca4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.977913] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Waiting for the task: (returnval){ [ 625.977913] env[61868]: value = "task-1315289" [ 625.977913] env[61868]: _type = "Task" [ 625.977913] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.984120] env[61868]: DEBUG nova.network.neutron [req-afa382e0-b841-4629-ba1a-3e7a5ead2a6f req-bdbf1ea3-10bf-47f2-ac62-7fb97d3d7f09 service nova] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.991163] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315289, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.127650] env[61868]: DEBUG nova.scheduler.client.report [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 626.256522] env[61868]: DEBUG nova.compute.manager [req-dcd00767-c355-47e4-8fc8-520b6f095d9e req-e1437968-3322-4928-a85b-28833d4790ee service nova] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Received event network-vif-deleted-d9e16fa5-610c-4053-8a20-1c33bf8d6324 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 626.389682] env[61868]: INFO nova.compute.manager [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] [instance: 5e812408-6f27-4b52-9564-4063f4526340] Took 1.05 seconds to deallocate network for instance. [ 626.447705] env[61868]: INFO nova.compute.manager [-] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Took 1.04 seconds to deallocate network for instance. [ 626.452180] env[61868]: DEBUG nova.compute.claims [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 626.452180] env[61868]: DEBUG oslo_concurrency.lockutils [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.488009] env[61868]: DEBUG oslo_concurrency.lockutils [req-afa382e0-b841-4629-ba1a-3e7a5ead2a6f req-bdbf1ea3-10bf-47f2-ac62-7fb97d3d7f09 service nova] Releasing lock "refresh_cache-7c2937c6-e6b7-4716-80fa-5538641bc8dc" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.488403] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315289, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.042644} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.488623] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] File moved {{(pid=61868) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 626.488812] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Cleaning up location [datastore2] vmware_temp/89d7d0cc-eef1-42bf-97df-1d0615b8fcd1 {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 626.488965] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Deleting the datastore file [datastore2] vmware_temp/89d7d0cc-eef1-42bf-97df-1d0615b8fcd1 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 626.489229] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a879013a-a343-493e-b342-994e3be90bf1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.496344] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Waiting for the task: (returnval){ [ 626.496344] env[61868]: value = "task-1315290" [ 626.496344] env[61868]: _type = "Task" [ 626.496344] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.504468] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315290, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.635188] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.621s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 626.635778] env[61868]: DEBUG nova.compute.manager [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 626.638455] env[61868]: DEBUG oslo_concurrency.lockutils [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.080s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.008248] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315290, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.025598} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.008248] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 627.008917] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ef4dc95-0d5f-4a04-8fff-29ff43f72099 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.016394] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Waiting for the task: (returnval){ [ 627.016394] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5295b76f-424f-9732-6444-54aac143712d" [ 627.016394] env[61868]: _type = "Task" [ 627.016394] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.025767] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5295b76f-424f-9732-6444-54aac143712d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.143839] env[61868]: DEBUG nova.compute.utils [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 627.149933] env[61868]: DEBUG nova.compute.manager [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 627.150333] env[61868]: DEBUG nova.network.neutron [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 627.233323] env[61868]: DEBUG nova.policy [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bf5777d27cff41479a2df2ea87168a70', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '02e2c51fbeb94fcfbf21cd0804f426a0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 627.428378] env[61868]: INFO nova.scheduler.client.report [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] Deleted allocations for instance 5e812408-6f27-4b52-9564-4063f4526340 [ 627.529990] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5295b76f-424f-9732-6444-54aac143712d, 'name': SearchDatastore_Task, 'duration_secs': 0.042372} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.533057] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 627.533598] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 7651f6d2-1825-46a1-9380-732b50abd3cc/7651f6d2-1825-46a1-9380-732b50abd3cc.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 627.534900] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-87a0c6b9-b799-4d97-8304-acf4662f2be4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.541855] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Waiting for the task: (returnval){ [ 627.541855] env[61868]: value = "task-1315291" [ 627.541855] env[61868]: _type = "Task" [ 627.541855] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.553534] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315291, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.654460] env[61868]: DEBUG nova.compute.manager [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 627.663760] env[61868]: DEBUG nova.network.neutron [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Successfully created port: 393c5ca8-99b5-438b-b7d1-7a6c503c8b5a {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 627.793191] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-518a2d64-d5ad-4403-adc2-5ea60c300ce4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.807802] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30aa52b4-8130-489d-a27a-2bc607537295 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.844426] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b505a28f-0446-439c-9780-6a11ac61d340 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.855440] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfac6356-0644-4dde-9ed1-aeb786893923 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.873605] env[61868]: DEBUG nova.compute.provider_tree [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 627.947081] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ca8fc3d2-1622-42a2-a5c3-1b5a36e13709 tempest-TenantUsagesTestJSON-1062371111 tempest-TenantUsagesTestJSON-1062371111-project-member] Lock "5e812408-6f27-4b52-9564-4063f4526340" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.591s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 628.055235] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315291, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.377429] env[61868]: DEBUG nova.scheduler.client.report [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 628.450226] env[61868]: DEBUG nova.compute.manager [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 628.521019] env[61868]: DEBUG nova.compute.manager [req-6ab2a418-9bf1-42a6-952e-7706160f52a3 req-6c2ca8bf-a5c7-45ac-a123-efe2caed53df service nova] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Received event network-changed-393c5ca8-99b5-438b-b7d1-7a6c503c8b5a {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 628.521019] env[61868]: DEBUG nova.compute.manager [req-6ab2a418-9bf1-42a6-952e-7706160f52a3 req-6c2ca8bf-a5c7-45ac-a123-efe2caed53df service nova] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Refreshing instance network info cache due to event network-changed-393c5ca8-99b5-438b-b7d1-7a6c503c8b5a. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 628.521148] env[61868]: DEBUG oslo_concurrency.lockutils [req-6ab2a418-9bf1-42a6-952e-7706160f52a3 req-6c2ca8bf-a5c7-45ac-a123-efe2caed53df service nova] Acquiring lock "refresh_cache-4c98c21e-2918-4b55-b2b5-db14630a36ef" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 628.521344] env[61868]: DEBUG oslo_concurrency.lockutils [req-6ab2a418-9bf1-42a6-952e-7706160f52a3 req-6c2ca8bf-a5c7-45ac-a123-efe2caed53df service nova] Acquired lock "refresh_cache-4c98c21e-2918-4b55-b2b5-db14630a36ef" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.521630] env[61868]: DEBUG nova.network.neutron [req-6ab2a418-9bf1-42a6-952e-7706160f52a3 req-6c2ca8bf-a5c7-45ac-a123-efe2caed53df service nova] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Refreshing network info cache for port 393c5ca8-99b5-438b-b7d1-7a6c503c8b5a {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 628.555587] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315291, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.647831} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.555857] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 7651f6d2-1825-46a1-9380-732b50abd3cc/7651f6d2-1825-46a1-9380-732b50abd3cc.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 628.556077] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 628.556321] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a045e503-bee2-4387-921f-94b076a9d4cd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.563369] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Waiting for the task: (returnval){ [ 628.563369] env[61868]: value = "task-1315292" [ 628.563369] env[61868]: _type = "Task" [ 628.563369] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.571231] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315292, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.666186] env[61868]: DEBUG nova.compute.manager [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 628.692727] env[61868]: DEBUG nova.virt.hardware [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 628.692971] env[61868]: DEBUG nova.virt.hardware [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 628.693159] env[61868]: DEBUG nova.virt.hardware [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 628.693350] env[61868]: DEBUG nova.virt.hardware [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 628.693496] env[61868]: DEBUG nova.virt.hardware [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 628.693640] env[61868]: DEBUG nova.virt.hardware [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 628.693846] env[61868]: DEBUG nova.virt.hardware [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 628.694009] env[61868]: DEBUG nova.virt.hardware [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 628.694432] env[61868]: DEBUG nova.virt.hardware [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 628.694861] env[61868]: DEBUG nova.virt.hardware [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 628.695190] env[61868]: DEBUG nova.virt.hardware [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 628.696214] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f8ebf39-3a97-4f02-8c30-33ced6ac7e0a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.707520] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc2ef896-13c8-46d1-abfc-d8ffd4389bd9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.774826] env[61868]: ERROR nova.compute.manager [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 393c5ca8-99b5-438b-b7d1-7a6c503c8b5a, please check neutron logs for more information. [ 628.774826] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 628.774826] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 628.774826] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 628.774826] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 628.774826] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 628.774826] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 628.774826] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 628.774826] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.774826] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 628.774826] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.774826] env[61868]: ERROR nova.compute.manager raise self.value [ 628.774826] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 628.774826] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 628.774826] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.774826] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 628.775282] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.775282] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 628.775282] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 393c5ca8-99b5-438b-b7d1-7a6c503c8b5a, please check neutron logs for more information. [ 628.775282] env[61868]: ERROR nova.compute.manager [ 628.775282] env[61868]: Traceback (most recent call last): [ 628.775282] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 628.775282] env[61868]: listener.cb(fileno) [ 628.775282] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 628.775282] env[61868]: result = function(*args, **kwargs) [ 628.775282] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 628.775282] env[61868]: return func(*args, **kwargs) [ 628.775282] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 628.775282] env[61868]: raise e [ 628.775282] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 628.775282] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 628.775282] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 628.775282] env[61868]: created_port_ids = self._update_ports_for_instance( [ 628.775282] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 628.775282] env[61868]: with excutils.save_and_reraise_exception(): [ 628.775282] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.775282] env[61868]: self.force_reraise() [ 628.775282] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.775282] env[61868]: raise self.value [ 628.775282] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 628.775282] env[61868]: updated_port = self._update_port( [ 628.775282] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.775282] env[61868]: _ensure_no_port_binding_failure(port) [ 628.775282] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.775282] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 628.775885] env[61868]: nova.exception.PortBindingFailed: Binding failed for port 393c5ca8-99b5-438b-b7d1-7a6c503c8b5a, please check neutron logs for more information. [ 628.775885] env[61868]: Removing descriptor: 16 [ 628.775885] env[61868]: ERROR nova.compute.manager [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 393c5ca8-99b5-438b-b7d1-7a6c503c8b5a, please check neutron logs for more information. [ 628.775885] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Traceback (most recent call last): [ 628.775885] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 628.775885] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] yield resources [ 628.775885] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 628.775885] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] self.driver.spawn(context, instance, image_meta, [ 628.775885] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 628.775885] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] self._vmops.spawn(context, instance, image_meta, injected_files, [ 628.775885] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 628.775885] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] vm_ref = self.build_virtual_machine(instance, [ 628.776156] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 628.776156] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] vif_infos = vmwarevif.get_vif_info(self._session, [ 628.776156] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 628.776156] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] for vif in network_info: [ 628.776156] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 628.776156] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] return self._sync_wrapper(fn, *args, **kwargs) [ 628.776156] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 628.776156] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] self.wait() [ 628.776156] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 628.776156] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] self[:] = self._gt.wait() [ 628.776156] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 628.776156] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] return self._exit_event.wait() [ 628.776156] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 628.776522] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] result = hub.switch() [ 628.776522] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 628.776522] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] return self.greenlet.switch() [ 628.776522] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 628.776522] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] result = function(*args, **kwargs) [ 628.776522] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 628.776522] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] return func(*args, **kwargs) [ 628.776522] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 628.776522] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] raise e [ 628.776522] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 628.776522] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] nwinfo = self.network_api.allocate_for_instance( [ 628.776522] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 628.776522] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] created_port_ids = self._update_ports_for_instance( [ 628.776825] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 628.776825] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] with excutils.save_and_reraise_exception(): [ 628.776825] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.776825] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] self.force_reraise() [ 628.776825] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.776825] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] raise self.value [ 628.776825] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 628.776825] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] updated_port = self._update_port( [ 628.776825] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.776825] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] _ensure_no_port_binding_failure(port) [ 628.776825] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.776825] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] raise exception.PortBindingFailed(port_id=port['id']) [ 628.777114] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] nova.exception.PortBindingFailed: Binding failed for port 393c5ca8-99b5-438b-b7d1-7a6c503c8b5a, please check neutron logs for more information. [ 628.777114] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] [ 628.777114] env[61868]: INFO nova.compute.manager [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Terminating instance [ 628.778376] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] Acquiring lock "refresh_cache-4c98c21e-2918-4b55-b2b5-db14630a36ef" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 628.885425] env[61868]: DEBUG oslo_concurrency.lockutils [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.247s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 628.886250] env[61868]: ERROR nova.compute.manager [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] [instance: e7439426-722a-4420-8607-903cf96dff06] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port af2beab0-e63a-4fd4-b1b8-999a8e0018d7, please check neutron logs for more information. [ 628.886250] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] Traceback (most recent call last): [ 628.886250] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 628.886250] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] self.driver.spawn(context, instance, image_meta, [ 628.886250] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 628.886250] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] self._vmops.spawn(context, instance, image_meta, injected_files, [ 628.886250] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 628.886250] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] vm_ref = self.build_virtual_machine(instance, [ 628.886250] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 628.886250] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] vif_infos = vmwarevif.get_vif_info(self._session, [ 628.886250] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 628.886606] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] for vif in network_info: [ 628.886606] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 628.886606] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] return self._sync_wrapper(fn, *args, **kwargs) [ 628.886606] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 628.886606] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] self.wait() [ 628.886606] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 628.886606] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] self[:] = self._gt.wait() [ 628.886606] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 628.886606] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] return self._exit_event.wait() [ 628.886606] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 628.886606] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] result = hub.switch() [ 628.886606] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 628.886606] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] return self.greenlet.switch() [ 628.886981] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 628.886981] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] result = function(*args, **kwargs) [ 628.886981] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 628.886981] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] return func(*args, **kwargs) [ 628.886981] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 628.886981] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] raise e [ 628.886981] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 628.886981] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] nwinfo = self.network_api.allocate_for_instance( [ 628.886981] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 628.886981] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] created_port_ids = self._update_ports_for_instance( [ 628.886981] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 628.886981] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] with excutils.save_and_reraise_exception(): [ 628.886981] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.887594] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] self.force_reraise() [ 628.887594] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.887594] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] raise self.value [ 628.887594] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 628.887594] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] updated_port = self._update_port( [ 628.887594] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.887594] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] _ensure_no_port_binding_failure(port) [ 628.887594] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.887594] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] raise exception.PortBindingFailed(port_id=port['id']) [ 628.887594] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] nova.exception.PortBindingFailed: Binding failed for port af2beab0-e63a-4fd4-b1b8-999a8e0018d7, please check neutron logs for more information. [ 628.887594] env[61868]: ERROR nova.compute.manager [instance: e7439426-722a-4420-8607-903cf96dff06] [ 628.887943] env[61868]: DEBUG nova.compute.utils [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] [instance: e7439426-722a-4420-8607-903cf96dff06] Binding failed for port af2beab0-e63a-4fd4-b1b8-999a8e0018d7, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 628.888050] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.475s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 628.892254] env[61868]: DEBUG nova.compute.manager [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] [instance: e7439426-722a-4420-8607-903cf96dff06] Build of instance e7439426-722a-4420-8607-903cf96dff06 was re-scheduled: Binding failed for port af2beab0-e63a-4fd4-b1b8-999a8e0018d7, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 628.892700] env[61868]: DEBUG nova.compute.manager [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] [instance: e7439426-722a-4420-8607-903cf96dff06] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 628.893017] env[61868]: DEBUG oslo_concurrency.lockutils [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] Acquiring lock "refresh_cache-e7439426-722a-4420-8607-903cf96dff06" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 628.893170] env[61868]: DEBUG oslo_concurrency.lockutils [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] Acquired lock "refresh_cache-e7439426-722a-4420-8607-903cf96dff06" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.893323] env[61868]: DEBUG nova.network.neutron [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] [instance: e7439426-722a-4420-8607-903cf96dff06] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 628.975215] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.045906] env[61868]: DEBUG nova.network.neutron [req-6ab2a418-9bf1-42a6-952e-7706160f52a3 req-6c2ca8bf-a5c7-45ac-a123-efe2caed53df service nova] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 629.074849] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315292, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068295} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.075234] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 629.076410] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b1de666-b114-43e9-9dbc-8391e5b81cba {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.103723] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Reconfiguring VM instance instance-00000011 to attach disk [datastore2] 7651f6d2-1825-46a1-9380-732b50abd3cc/7651f6d2-1825-46a1-9380-732b50abd3cc.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 629.104228] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-75aab23e-ffe0-4e76-af49-7bf3bbfa2683 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.127999] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Waiting for the task: (returnval){ [ 629.127999] env[61868]: value = "task-1315293" [ 629.127999] env[61868]: _type = "Task" [ 629.127999] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.137167] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315293, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.220371] env[61868]: DEBUG nova.network.neutron [req-6ab2a418-9bf1-42a6-952e-7706160f52a3 req-6c2ca8bf-a5c7-45ac-a123-efe2caed53df service nova] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.413241] env[61868]: DEBUG nova.network.neutron [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] [instance: e7439426-722a-4420-8607-903cf96dff06] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 629.537784] env[61868]: DEBUG nova.network.neutron [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] [instance: e7439426-722a-4420-8607-903cf96dff06] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.639171] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315293, 'name': ReconfigVM_Task, 'duration_secs': 0.297332} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.639684] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Reconfigured VM instance instance-00000011 to attach disk [datastore2] 7651f6d2-1825-46a1-9380-732b50abd3cc/7651f6d2-1825-46a1-9380-732b50abd3cc.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 629.640059] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-77482d62-499a-4e08-b9b2-89fafe3d5dc2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.646182] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Waiting for the task: (returnval){ [ 629.646182] env[61868]: value = "task-1315294" [ 629.646182] env[61868]: _type = "Task" [ 629.646182] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.656257] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315294, 'name': Rename_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.725719] env[61868]: DEBUG oslo_concurrency.lockutils [req-6ab2a418-9bf1-42a6-952e-7706160f52a3 req-6c2ca8bf-a5c7-45ac-a123-efe2caed53df service nova] Releasing lock "refresh_cache-4c98c21e-2918-4b55-b2b5-db14630a36ef" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.726096] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] Acquired lock "refresh_cache-4c98c21e-2918-4b55-b2b5-db14630a36ef" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.729021] env[61868]: DEBUG nova.network.neutron [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 629.876202] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-423bce0d-d7cb-4936-8e90-baac85a40e24 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.884842] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9bebde9-061f-4b0f-8ae0-ae7fd47b78e6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.917887] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a4b7812-4433-4970-aac3-255a3511b162 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.928239] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fbe06c1-a7a5-41b2-b128-bda6279c157e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.949017] env[61868]: DEBUG nova.compute.provider_tree [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 630.046017] env[61868]: DEBUG oslo_concurrency.lockutils [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] Releasing lock "refresh_cache-e7439426-722a-4420-8607-903cf96dff06" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 630.046017] env[61868]: DEBUG nova.compute.manager [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 630.046017] env[61868]: DEBUG nova.compute.manager [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] [instance: e7439426-722a-4420-8607-903cf96dff06] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 630.046017] env[61868]: DEBUG nova.network.neutron [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] [instance: e7439426-722a-4420-8607-903cf96dff06] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 630.069163] env[61868]: DEBUG nova.network.neutron [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] [instance: e7439426-722a-4420-8607-903cf96dff06] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 630.159040] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315294, 'name': Rename_Task, 'duration_secs': 0.129908} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.159378] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 630.159655] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-883caea7-4486-4180-8e86-e890abbb216d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.166712] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Waiting for the task: (returnval){ [ 630.166712] env[61868]: value = "task-1315295" [ 630.166712] env[61868]: _type = "Task" [ 630.166712] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.175653] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315295, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.247249] env[61868]: DEBUG nova.network.neutron [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 630.328114] env[61868]: DEBUG nova.network.neutron [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.451256] env[61868]: DEBUG nova.scheduler.client.report [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 630.459257] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Acquiring lock "4de32ee6-e205-435b-a75e-94c3121048e3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.459257] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Lock "4de32ee6-e205-435b-a75e-94c3121048e3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.573884] env[61868]: DEBUG nova.network.neutron [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] [instance: e7439426-722a-4420-8607-903cf96dff06] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.675925] env[61868]: DEBUG oslo_vmware.api [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315295, 'name': PowerOnVM_Task, 'duration_secs': 0.465042} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.677193] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 630.677398] env[61868]: DEBUG nova.compute.manager [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 630.678572] env[61868]: DEBUG nova.compute.manager [req-057ed646-2000-4c2c-997a-bbe8be2f960f req-fbf1eb22-43d6-4f7e-8e4c-0c430ef992af service nova] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Received event network-vif-deleted-393c5ca8-99b5-438b-b7d1-7a6c503c8b5a {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 630.679329] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c180cc25-f97f-46c6-9653-93fad72f416b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.833371] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] Releasing lock "refresh_cache-4c98c21e-2918-4b55-b2b5-db14630a36ef" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 630.833752] env[61868]: DEBUG nova.compute.manager [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 630.833945] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 630.834254] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8cc93479-da5a-499e-bbff-3f5a90ccc33b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.843047] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3332d22e-b879-49d0-9807-afb2f80bf871 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.865096] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4c98c21e-2918-4b55-b2b5-db14630a36ef could not be found. [ 630.865333] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 630.865514] env[61868]: INFO nova.compute.manager [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Took 0.03 seconds to destroy the instance on the hypervisor. [ 630.865754] env[61868]: DEBUG oslo.service.loopingcall [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 630.865976] env[61868]: DEBUG nova.compute.manager [-] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 630.866083] env[61868]: DEBUG nova.network.neutron [-] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 630.882054] env[61868]: DEBUG nova.network.neutron [-] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 630.963609] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.071s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.963609] env[61868]: ERROR nova.compute.manager [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f0536e29-ba69-4dc1-a336-c44893ce3afc, please check neutron logs for more information. [ 630.963609] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Traceback (most recent call last): [ 630.963609] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 630.963609] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] self.driver.spawn(context, instance, image_meta, [ 630.963609] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 630.963609] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] self._vmops.spawn(context, instance, image_meta, injected_files, [ 630.963609] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 630.963609] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] vm_ref = self.build_virtual_machine(instance, [ 630.964780] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 630.964780] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] vif_infos = vmwarevif.get_vif_info(self._session, [ 630.964780] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 630.964780] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] for vif in network_info: [ 630.964780] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 630.964780] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] return self._sync_wrapper(fn, *args, **kwargs) [ 630.964780] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 630.964780] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] self.wait() [ 630.964780] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 630.964780] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] self[:] = self._gt.wait() [ 630.964780] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 630.964780] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] return self._exit_event.wait() [ 630.964780] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 630.965103] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] result = hub.switch() [ 630.965103] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 630.965103] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] return self.greenlet.switch() [ 630.965103] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 630.965103] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] result = function(*args, **kwargs) [ 630.965103] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 630.965103] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] return func(*args, **kwargs) [ 630.965103] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 630.965103] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] raise e [ 630.965103] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 630.965103] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] nwinfo = self.network_api.allocate_for_instance( [ 630.965103] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 630.965103] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] created_port_ids = self._update_ports_for_instance( [ 630.965393] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 630.965393] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] with excutils.save_and_reraise_exception(): [ 630.965393] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 630.965393] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] self.force_reraise() [ 630.965393] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 630.965393] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] raise self.value [ 630.965393] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 630.965393] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] updated_port = self._update_port( [ 630.965393] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 630.965393] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] _ensure_no_port_binding_failure(port) [ 630.965393] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 630.965393] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] raise exception.PortBindingFailed(port_id=port['id']) [ 630.965652] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] nova.exception.PortBindingFailed: Binding failed for port f0536e29-ba69-4dc1-a336-c44893ce3afc, please check neutron logs for more information. [ 630.965652] env[61868]: ERROR nova.compute.manager [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] [ 630.965652] env[61868]: DEBUG nova.compute.utils [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Binding failed for port f0536e29-ba69-4dc1-a336-c44893ce3afc, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 630.965652] env[61868]: DEBUG oslo_concurrency.lockutils [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.552s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.968355] env[61868]: DEBUG nova.compute.manager [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Build of instance a93e681e-1cb1-4025-8725-4c6664d78c96 was re-scheduled: Binding failed for port f0536e29-ba69-4dc1-a336-c44893ce3afc, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 630.968705] env[61868]: DEBUG nova.compute.manager [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 630.969042] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] Acquiring lock "refresh_cache-a93e681e-1cb1-4025-8725-4c6664d78c96" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 630.969265] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] Acquired lock "refresh_cache-a93e681e-1cb1-4025-8725-4c6664d78c96" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.969462] env[61868]: DEBUG nova.network.neutron [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 631.076977] env[61868]: INFO nova.compute.manager [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] [instance: e7439426-722a-4420-8607-903cf96dff06] Took 1.03 seconds to deallocate network for instance. [ 631.121509] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] Acquiring lock "a0fda7cf-21b6-4a22-bf56-09e848b4dc9e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.121799] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] Lock "a0fda7cf-21b6-4a22-bf56-09e848b4dc9e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.198503] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.384225] env[61868]: DEBUG nova.network.neutron [-] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.496179] env[61868]: DEBUG nova.network.neutron [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 631.532891] env[61868]: INFO nova.compute.manager [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Rebuilding instance [ 631.587694] env[61868]: DEBUG nova.compute.manager [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 631.588914] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62aba87a-4f0a-40ed-b585-8d7228a41b83 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.599907] env[61868]: DEBUG nova.network.neutron [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.886913] env[61868]: INFO nova.compute.manager [-] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Took 1.02 seconds to deallocate network for instance. [ 631.890506] env[61868]: DEBUG nova.compute.claims [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 631.890685] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.961822] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-071cf0b7-4306-4b93-85af-09df452a392f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.968632] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11d4a6dd-efea-41fe-bfc2-00612c07db20 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.998545] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e02752e-03c2-4e6a-b528-9448190f5885 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.006585] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffde7f3f-89e5-4e05-a677-940391259499 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.021615] env[61868]: DEBUG nova.compute.provider_tree [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 632.107043] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 632.108146] env[61868]: INFO nova.scheduler.client.report [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] Deleted allocations for instance e7439426-722a-4420-8607-903cf96dff06 [ 632.114142] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] Releasing lock "refresh_cache-a93e681e-1cb1-4025-8725-4c6664d78c96" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.114356] env[61868]: DEBUG nova.compute.manager [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 632.114535] env[61868]: DEBUG nova.compute.manager [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 632.114699] env[61868]: DEBUG nova.network.neutron [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 632.116350] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-708a6139-f381-4a0f-93e8-d7b49865b947 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.123994] env[61868]: DEBUG oslo_vmware.api [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Waiting for the task: (returnval){ [ 632.123994] env[61868]: value = "task-1315296" [ 632.123994] env[61868]: _type = "Task" [ 632.123994] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.132326] env[61868]: DEBUG oslo_vmware.api [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Task: {'id': task-1315296, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.134811] env[61868]: DEBUG nova.network.neutron [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 632.525324] env[61868]: DEBUG nova.scheduler.client.report [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 632.620843] env[61868]: DEBUG oslo_concurrency.lockutils [None req-83789ec3-3c17-43de-99fc-60b347d90471 tempest-FloatingIPsAssociationTestJSON-273417552 tempest-FloatingIPsAssociationTestJSON-273417552-project-member] Lock "e7439426-722a-4420-8607-903cf96dff06" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.474s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 632.633926] env[61868]: DEBUG oslo_vmware.api [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Task: {'id': task-1315296, 'name': PowerOffVM_Task, 'duration_secs': 0.184019} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 632.634198] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 632.634414] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 632.635189] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8173627-fe61-44ed-9227-5c0a78a9443c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.638034] env[61868]: DEBUG nova.network.neutron [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.643350] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 632.643609] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c9f3bd54-c16e-48b7-a339-b465d245b68f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.667097] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 632.667338] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Deleting contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 632.667520] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Deleting the datastore file [datastore2] 7651f6d2-1825-46a1-9380-732b50abd3cc {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 632.667895] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9d53fdc4-ba28-4474-82ad-a90714fcb664 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.675100] env[61868]: DEBUG oslo_vmware.api [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Waiting for the task: (returnval){ [ 632.675100] env[61868]: value = "task-1315298" [ 632.675100] env[61868]: _type = "Task" [ 632.675100] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.683558] env[61868]: DEBUG oslo_vmware.api [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Task: {'id': task-1315298, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.031848] env[61868]: DEBUG oslo_concurrency.lockutils [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.067s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 633.032383] env[61868]: ERROR nova.compute.manager [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0d931638-71c7-4f34-aff9-273312a018e6, please check neutron logs for more information. [ 633.032383] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Traceback (most recent call last): [ 633.032383] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 633.032383] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] self.driver.spawn(context, instance, image_meta, [ 633.032383] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 633.032383] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] self._vmops.spawn(context, instance, image_meta, injected_files, [ 633.032383] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 633.032383] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] vm_ref = self.build_virtual_machine(instance, [ 633.032383] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 633.032383] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] vif_infos = vmwarevif.get_vif_info(self._session, [ 633.032383] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 633.032704] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] for vif in network_info: [ 633.032704] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 633.032704] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] return self._sync_wrapper(fn, *args, **kwargs) [ 633.032704] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 633.032704] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] self.wait() [ 633.032704] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 633.032704] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] self[:] = self._gt.wait() [ 633.032704] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 633.032704] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] return self._exit_event.wait() [ 633.032704] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 633.032704] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] result = hub.switch() [ 633.032704] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 633.032704] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] return self.greenlet.switch() [ 633.033014] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 633.033014] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] result = function(*args, **kwargs) [ 633.033014] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 633.033014] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] return func(*args, **kwargs) [ 633.033014] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 633.033014] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] raise e [ 633.033014] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 633.033014] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] nwinfo = self.network_api.allocate_for_instance( [ 633.033014] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 633.033014] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] created_port_ids = self._update_ports_for_instance( [ 633.033014] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 633.033014] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] with excutils.save_and_reraise_exception(): [ 633.033014] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.033327] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] self.force_reraise() [ 633.033327] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.033327] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] raise self.value [ 633.033327] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 633.033327] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] updated_port = self._update_port( [ 633.033327] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.033327] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] _ensure_no_port_binding_failure(port) [ 633.033327] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.033327] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] raise exception.PortBindingFailed(port_id=port['id']) [ 633.033327] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] nova.exception.PortBindingFailed: Binding failed for port 0d931638-71c7-4f34-aff9-273312a018e6, please check neutron logs for more information. [ 633.033327] env[61868]: ERROR nova.compute.manager [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] [ 633.033938] env[61868]: DEBUG nova.compute.utils [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Binding failed for port 0d931638-71c7-4f34-aff9-273312a018e6, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 633.035300] env[61868]: DEBUG oslo_concurrency.lockutils [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.214s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.038532] env[61868]: DEBUG nova.compute.manager [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Build of instance 448b8246-8d41-4154-bb99-b3c66d8e9475 was re-scheduled: Binding failed for port 0d931638-71c7-4f34-aff9-273312a018e6, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 633.039045] env[61868]: DEBUG nova.compute.manager [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 633.039334] env[61868]: DEBUG oslo_concurrency.lockutils [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] Acquiring lock "refresh_cache-448b8246-8d41-4154-bb99-b3c66d8e9475" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 633.039536] env[61868]: DEBUG oslo_concurrency.lockutils [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] Acquired lock "refresh_cache-448b8246-8d41-4154-bb99-b3c66d8e9475" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.039747] env[61868]: DEBUG nova.network.neutron [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 633.123684] env[61868]: DEBUG nova.compute.manager [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 633.140962] env[61868]: INFO nova.compute.manager [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] [instance: a93e681e-1cb1-4025-8725-4c6664d78c96] Took 1.03 seconds to deallocate network for instance. [ 633.189203] env[61868]: DEBUG oslo_vmware.api [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Task: {'id': task-1315298, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.09406} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.189537] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 633.189652] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Deleted contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 633.189848] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 633.579581] env[61868]: DEBUG nova.network.neutron [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 633.647233] env[61868]: DEBUG oslo_concurrency.lockutils [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.713145] env[61868]: DEBUG nova.network.neutron [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.016227] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2db72ed4-63bd-4357-b853-966fd56db766 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.025265] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-251de22b-9e62-4827-9c8a-1d0c010cd1dc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.054870] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ef620b6-f6b8-49d2-ae3a-a6a7f1fa4e91 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.063008] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50864644-1463-49e7-84a7-598bac40fa5e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.076328] env[61868]: DEBUG nova.compute.provider_tree [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 634.173316] env[61868]: INFO nova.scheduler.client.report [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] Deleted allocations for instance a93e681e-1cb1-4025-8725-4c6664d78c96 [ 634.218227] env[61868]: DEBUG oslo_concurrency.lockutils [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] Releasing lock "refresh_cache-448b8246-8d41-4154-bb99-b3c66d8e9475" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.218719] env[61868]: DEBUG nova.compute.manager [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 634.218719] env[61868]: DEBUG nova.compute.manager [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 634.218813] env[61868]: DEBUG nova.network.neutron [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 634.229510] env[61868]: DEBUG nova.virt.hardware [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 634.229510] env[61868]: DEBUG nova.virt.hardware [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 634.229998] env[61868]: DEBUG nova.virt.hardware [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 634.233324] env[61868]: DEBUG nova.virt.hardware [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 634.233324] env[61868]: DEBUG nova.virt.hardware [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 634.233324] env[61868]: DEBUG nova.virt.hardware [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 634.233324] env[61868]: DEBUG nova.virt.hardware [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 634.233324] env[61868]: DEBUG nova.virt.hardware [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 634.233577] env[61868]: DEBUG nova.virt.hardware [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 634.233577] env[61868]: DEBUG nova.virt.hardware [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 634.233577] env[61868]: DEBUG nova.virt.hardware [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 634.233706] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4847dad-5347-4dcb-a590-ee80f824952f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.238444] env[61868]: DEBUG nova.network.neutron [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 634.242881] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a2fd613-caa8-4c5b-8dbb-9712565813e9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.258945] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Instance VIF info [] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 634.264927] env[61868]: DEBUG oslo.service.loopingcall [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 634.265205] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 634.265489] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bef8f358-7659-4277-81f9-1291b4c9bb8d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.282417] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 634.282417] env[61868]: value = "task-1315299" [ 634.282417] env[61868]: _type = "Task" [ 634.282417] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.290181] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315299, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.579979] env[61868]: DEBUG nova.scheduler.client.report [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 634.682704] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5ef68f1b-44ee-4b6e-b9a7-3dcfd2fe210a tempest-ImagesNegativeTestJSON-1132951458 tempest-ImagesNegativeTestJSON-1132951458-project-member] Lock "a93e681e-1cb1-4025-8725-4c6664d78c96" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 75.687s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 634.744114] env[61868]: DEBUG nova.network.neutron [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.792746] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315299, 'name': CreateVM_Task, 'duration_secs': 0.248201} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.792896] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 634.793326] env[61868]: DEBUG oslo_concurrency.lockutils [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 634.793479] env[61868]: DEBUG oslo_concurrency.lockutils [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.793782] env[61868]: DEBUG oslo_concurrency.lockutils [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 634.794034] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-639dda4f-4eb2-4145-9fd0-d187701294b8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.798392] env[61868]: DEBUG oslo_vmware.api [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Waiting for the task: (returnval){ [ 634.798392] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]523e44d3-5bb8-90f2-513e-c3be3853e464" [ 634.798392] env[61868]: _type = "Task" [ 634.798392] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.805732] env[61868]: DEBUG oslo_vmware.api [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]523e44d3-5bb8-90f2-513e-c3be3853e464, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.084528] env[61868]: DEBUG oslo_concurrency.lockutils [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.049s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 635.085187] env[61868]: ERROR nova.compute.manager [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8354b128-a713-44f1-9205-d5dc691119b0, please check neutron logs for more information. [ 635.085187] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Traceback (most recent call last): [ 635.085187] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 635.085187] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] self.driver.spawn(context, instance, image_meta, [ 635.085187] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 635.085187] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 635.085187] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 635.085187] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] vm_ref = self.build_virtual_machine(instance, [ 635.085187] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 635.085187] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] vif_infos = vmwarevif.get_vif_info(self._session, [ 635.085187] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 635.085445] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] for vif in network_info: [ 635.085445] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 635.085445] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] return self._sync_wrapper(fn, *args, **kwargs) [ 635.085445] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 635.085445] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] self.wait() [ 635.085445] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 635.085445] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] self[:] = self._gt.wait() [ 635.085445] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 635.085445] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] return self._exit_event.wait() [ 635.085445] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 635.085445] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] result = hub.switch() [ 635.085445] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 635.085445] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] return self.greenlet.switch() [ 635.085725] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 635.085725] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] result = function(*args, **kwargs) [ 635.085725] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 635.085725] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] return func(*args, **kwargs) [ 635.085725] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 635.085725] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] raise e [ 635.085725] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.085725] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] nwinfo = self.network_api.allocate_for_instance( [ 635.085725] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 635.085725] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] created_port_ids = self._update_ports_for_instance( [ 635.085725] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 635.085725] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] with excutils.save_and_reraise_exception(): [ 635.085725] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.086014] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] self.force_reraise() [ 635.086014] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.086014] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] raise self.value [ 635.086014] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 635.086014] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] updated_port = self._update_port( [ 635.086014] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.086014] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] _ensure_no_port_binding_failure(port) [ 635.086014] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.086014] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] raise exception.PortBindingFailed(port_id=port['id']) [ 635.086014] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] nova.exception.PortBindingFailed: Binding failed for port 8354b128-a713-44f1-9205-d5dc691119b0, please check neutron logs for more information. [ 635.086014] env[61868]: ERROR nova.compute.manager [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] [ 635.086261] env[61868]: DEBUG nova.compute.utils [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Binding failed for port 8354b128-a713-44f1-9205-d5dc691119b0, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 635.087823] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.893s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 635.089838] env[61868]: INFO nova.compute.claims [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 635.096460] env[61868]: DEBUG nova.compute.manager [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Build of instance 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d was re-scheduled: Binding failed for port 8354b128-a713-44f1-9205-d5dc691119b0, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 635.096989] env[61868]: DEBUG nova.compute.manager [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 635.097315] env[61868]: DEBUG oslo_concurrency.lockutils [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] Acquiring lock "refresh_cache-2f9deec0-e0ad-4cb4-9868-9aa53cdf602d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 635.097514] env[61868]: DEBUG oslo_concurrency.lockutils [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] Acquired lock "refresh_cache-2f9deec0-e0ad-4cb4-9868-9aa53cdf602d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.097734] env[61868]: DEBUG nova.network.neutron [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 635.184981] env[61868]: DEBUG nova.compute.manager [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 635.249819] env[61868]: INFO nova.compute.manager [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] [instance: 448b8246-8d41-4154-bb99-b3c66d8e9475] Took 1.03 seconds to deallocate network for instance. [ 635.316067] env[61868]: DEBUG oslo_vmware.api [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]523e44d3-5bb8-90f2-513e-c3be3853e464, 'name': SearchDatastore_Task, 'duration_secs': 0.016714} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.316067] env[61868]: DEBUG oslo_concurrency.lockutils [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 635.316067] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 635.316067] env[61868]: DEBUG oslo_concurrency.lockutils [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 635.316314] env[61868]: DEBUG oslo_concurrency.lockutils [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.316314] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 635.316314] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-775c56ae-c19d-43d7-8da3-c21fb10a710c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.322621] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 635.322952] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 635.323968] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cedc400f-07f5-49ef-a865-cc6c6245317e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.329261] env[61868]: DEBUG oslo_vmware.api [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Waiting for the task: (returnval){ [ 635.329261] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52db2da1-ba0f-4d11-b256-e17484d70f2a" [ 635.329261] env[61868]: _type = "Task" [ 635.329261] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.337791] env[61868]: DEBUG oslo_vmware.api [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52db2da1-ba0f-4d11-b256-e17484d70f2a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.631170] env[61868]: DEBUG nova.network.neutron [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 635.711188] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.744505] env[61868]: DEBUG nova.network.neutron [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.842425] env[61868]: DEBUG oslo_vmware.api [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52db2da1-ba0f-4d11-b256-e17484d70f2a, 'name': SearchDatastore_Task, 'duration_secs': 0.007996} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.843573] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bdf14c8e-e834-480e-bb83-3d834fafe4e9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.848951] env[61868]: DEBUG oslo_vmware.api [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Waiting for the task: (returnval){ [ 635.848951] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52381f2f-4b2f-5dd8-296c-a41d39a75287" [ 635.848951] env[61868]: _type = "Task" [ 635.848951] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.858845] env[61868]: DEBUG oslo_vmware.api [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52381f2f-4b2f-5dd8-296c-a41d39a75287, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.248495] env[61868]: DEBUG oslo_concurrency.lockutils [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] Releasing lock "refresh_cache-2f9deec0-e0ad-4cb4-9868-9aa53cdf602d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.249734] env[61868]: DEBUG nova.compute.manager [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 636.250043] env[61868]: DEBUG nova.compute.manager [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 636.250278] env[61868]: DEBUG nova.network.neutron [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 636.270767] env[61868]: DEBUG nova.network.neutron [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 636.293505] env[61868]: INFO nova.scheduler.client.report [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] Deleted allocations for instance 448b8246-8d41-4154-bb99-b3c66d8e9475 [ 636.366394] env[61868]: DEBUG oslo_vmware.api [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52381f2f-4b2f-5dd8-296c-a41d39a75287, 'name': SearchDatastore_Task, 'duration_secs': 0.008539} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.366394] env[61868]: DEBUG oslo_concurrency.lockutils [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.366394] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 7651f6d2-1825-46a1-9380-732b50abd3cc/7651f6d2-1825-46a1-9380-732b50abd3cc.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 636.366714] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4e89a37e-e2a8-4b93-a9e6-e5d3065c3885 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.377158] env[61868]: DEBUG oslo_vmware.api [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Waiting for the task: (returnval){ [ 636.377158] env[61868]: value = "task-1315300" [ 636.377158] env[61868]: _type = "Task" [ 636.377158] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.384511] env[61868]: DEBUG oslo_vmware.api [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Task: {'id': task-1315300, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.633696] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24bc4e03-f4bd-46d5-be5a-18991441a8d5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.647540] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f711052-adaa-49c4-82e0-b6261592d111 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.684873] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2aa2e05-7fd3-4706-b3ea-f95e29d5aeb2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.694035] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bcced24-e052-4992-aeb9-23c2c1facceb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.711572] env[61868]: DEBUG nova.compute.provider_tree [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 636.774165] env[61868]: DEBUG nova.network.neutron [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.808920] env[61868]: DEBUG oslo_concurrency.lockutils [None req-54ac10ea-b425-4ba7-af18-fd1aab3d244f tempest-ServersWithSpecificFlavorTestJSON-973145047 tempest-ServersWithSpecificFlavorTestJSON-973145047-project-member] Lock "448b8246-8d41-4154-bb99-b3c66d8e9475" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 76.270s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 636.886292] env[61868]: DEBUG oslo_vmware.api [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Task: {'id': task-1315300, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.217054] env[61868]: DEBUG nova.scheduler.client.report [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 637.276927] env[61868]: INFO nova.compute.manager [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] [instance: 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d] Took 1.03 seconds to deallocate network for instance. [ 637.314153] env[61868]: DEBUG nova.compute.manager [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 637.386532] env[61868]: DEBUG oslo_vmware.api [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Task: {'id': task-1315300, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.534754} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 637.386775] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 7651f6d2-1825-46a1-9380-732b50abd3cc/7651f6d2-1825-46a1-9380-732b50abd3cc.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 637.387032] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 637.387283] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7c06402a-b783-4cb7-a7cf-609ffc6bed92 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.398171] env[61868]: DEBUG oslo_vmware.api [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Waiting for the task: (returnval){ [ 637.398171] env[61868]: value = "task-1315301" [ 637.398171] env[61868]: _type = "Task" [ 637.398171] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.406459] env[61868]: DEBUG oslo_vmware.api [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Task: {'id': task-1315301, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.724109] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.636s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 637.724703] env[61868]: DEBUG nova.compute.manager [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 637.727250] env[61868]: DEBUG oslo_concurrency.lockutils [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.031s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.843211] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.908824] env[61868]: DEBUG oslo_vmware.api [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Task: {'id': task-1315301, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066069} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 637.909165] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 637.909881] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01612aa7-6d67-40d1-ba4b-389abc8bbbb3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.929939] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Reconfiguring VM instance instance-00000011 to attach disk [datastore2] 7651f6d2-1825-46a1-9380-732b50abd3cc/7651f6d2-1825-46a1-9380-732b50abd3cc.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 637.932061] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e536d0be-9f0b-4681-8810-c1658b875bd3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.951792] env[61868]: DEBUG oslo_vmware.api [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Waiting for the task: (returnval){ [ 637.951792] env[61868]: value = "task-1315302" [ 637.951792] env[61868]: _type = "Task" [ 637.951792] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.960106] env[61868]: DEBUG oslo_vmware.api [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Task: {'id': task-1315302, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.232472] env[61868]: DEBUG nova.compute.utils [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 638.238592] env[61868]: DEBUG nova.compute.manager [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 638.238781] env[61868]: DEBUG nova.network.neutron [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 638.298754] env[61868]: DEBUG nova.policy [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '52e3605b4dff4ab8ae647a73df70b72a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cc31ff1aa47c457c9092cbe77f585d77', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 638.323519] env[61868]: INFO nova.scheduler.client.report [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] Deleted allocations for instance 2f9deec0-e0ad-4cb4-9868-9aa53cdf602d [ 638.463210] env[61868]: DEBUG oslo_vmware.api [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Task: {'id': task-1315302, 'name': ReconfigVM_Task, 'duration_secs': 0.288637} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.463627] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Reconfigured VM instance instance-00000011 to attach disk [datastore2] 7651f6d2-1825-46a1-9380-732b50abd3cc/7651f6d2-1825-46a1-9380-732b50abd3cc.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 638.464180] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-38af51f3-57c1-4a1a-9971-5c3f48b504e1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.472492] env[61868]: DEBUG oslo_vmware.api [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Waiting for the task: (returnval){ [ 638.472492] env[61868]: value = "task-1315303" [ 638.472492] env[61868]: _type = "Task" [ 638.472492] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.480342] env[61868]: DEBUG oslo_vmware.api [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Task: {'id': task-1315303, 'name': Rename_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.671761] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45fa090f-0c2e-49bf-96bf-b5981e3a4b7e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.682374] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-518c93f7-8de8-441f-9a3f-32e62df99534 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.714994] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81a85f8d-5229-4ea0-8115-14236ce49947 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.725275] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f50f4a0a-c707-48df-a9d7-139dfa38f140 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.738947] env[61868]: DEBUG nova.compute.manager [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 638.741777] env[61868]: DEBUG nova.compute.provider_tree [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 638.800566] env[61868]: DEBUG nova.network.neutron [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Successfully created port: a4c26f3e-a6e6-45b9-b449-bf07b77ca99b {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 638.840226] env[61868]: DEBUG oslo_concurrency.lockutils [None req-591c9282-74df-46bf-9c5f-496f7af16985 tempest-ImagesOneServerTestJSON-706025484 tempest-ImagesOneServerTestJSON-706025484-project-member] Lock "2f9deec0-e0ad-4cb4-9868-9aa53cdf602d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 76.863s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 638.984446] env[61868]: DEBUG oslo_vmware.api [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Task: {'id': task-1315303, 'name': Rename_Task, 'duration_secs': 0.130829} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.984710] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 638.984946] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-945e15d4-3117-4445-b95f-4be8a5d032d0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.991451] env[61868]: DEBUG oslo_vmware.api [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Waiting for the task: (returnval){ [ 638.991451] env[61868]: value = "task-1315304" [ 638.991451] env[61868]: _type = "Task" [ 638.991451] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.998984] env[61868]: DEBUG oslo_vmware.api [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Task: {'id': task-1315304, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.254685] env[61868]: DEBUG nova.scheduler.client.report [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 639.343757] env[61868]: DEBUG nova.compute.manager [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 639.420992] env[61868]: DEBUG oslo_concurrency.lockutils [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "ebb09f50-50b8-49bb-b366-5b8906605699" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.420992] env[61868]: DEBUG oslo_concurrency.lockutils [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "ebb09f50-50b8-49bb-b366-5b8906605699" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.502086] env[61868]: DEBUG oslo_vmware.api [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Task: {'id': task-1315304, 'name': PowerOnVM_Task, 'duration_secs': 0.435537} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.502391] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 639.502587] env[61868]: DEBUG nova.compute.manager [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 639.504169] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71a7bbd4-fd6d-4df7-a783-d157e39a218e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.756105] env[61868]: DEBUG nova.compute.manager [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 639.760273] env[61868]: DEBUG oslo_concurrency.lockutils [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.030s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 639.760273] env[61868]: ERROR nova.compute.manager [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a0af3a45-2562-432e-9a98-d8ed20d76195, please check neutron logs for more information. [ 639.760273] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Traceback (most recent call last): [ 639.760273] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 639.760273] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] self.driver.spawn(context, instance, image_meta, [ 639.760273] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 639.760273] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] self._vmops.spawn(context, instance, image_meta, injected_files, [ 639.760273] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 639.760273] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] vm_ref = self.build_virtual_machine(instance, [ 639.760503] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 639.760503] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] vif_infos = vmwarevif.get_vif_info(self._session, [ 639.760503] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 639.760503] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] for vif in network_info: [ 639.760503] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 639.760503] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] return self._sync_wrapper(fn, *args, **kwargs) [ 639.760503] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 639.760503] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] self.wait() [ 639.760503] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 639.760503] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] self[:] = self._gt.wait() [ 639.760503] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 639.760503] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] return self._exit_event.wait() [ 639.760503] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 639.760785] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] result = hub.switch() [ 639.760785] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 639.760785] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] return self.greenlet.switch() [ 639.760785] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 639.760785] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] result = function(*args, **kwargs) [ 639.760785] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 639.760785] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] return func(*args, **kwargs) [ 639.760785] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 639.760785] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] raise e [ 639.760785] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 639.760785] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] nwinfo = self.network_api.allocate_for_instance( [ 639.760785] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 639.760785] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] created_port_ids = self._update_ports_for_instance( [ 639.761105] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 639.761105] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] with excutils.save_and_reraise_exception(): [ 639.761105] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 639.761105] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] self.force_reraise() [ 639.761105] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 639.761105] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] raise self.value [ 639.761105] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 639.761105] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] updated_port = self._update_port( [ 639.761105] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 639.761105] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] _ensure_no_port_binding_failure(port) [ 639.761105] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 639.761105] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] raise exception.PortBindingFailed(port_id=port['id']) [ 639.761366] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] nova.exception.PortBindingFailed: Binding failed for port a0af3a45-2562-432e-9a98-d8ed20d76195, please check neutron logs for more information. [ 639.761366] env[61868]: ERROR nova.compute.manager [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] [ 639.761366] env[61868]: DEBUG nova.compute.utils [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Binding failed for port a0af3a45-2562-432e-9a98-d8ed20d76195, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 639.761366] env[61868]: DEBUG oslo_concurrency.lockutils [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.666s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.764663] env[61868]: INFO nova.compute.claims [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 639.769268] env[61868]: DEBUG nova.compute.manager [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Build of instance 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431 was re-scheduled: Binding failed for port a0af3a45-2562-432e-9a98-d8ed20d76195, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 639.769808] env[61868]: DEBUG nova.compute.manager [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 639.769808] env[61868]: DEBUG oslo_concurrency.lockutils [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Acquiring lock "refresh_cache-8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 639.770095] env[61868]: DEBUG oslo_concurrency.lockutils [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Acquired lock "refresh_cache-8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.770225] env[61868]: DEBUG nova.network.neutron [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 639.805406] env[61868]: DEBUG nova.virt.hardware [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 639.807123] env[61868]: DEBUG nova.virt.hardware [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 639.807123] env[61868]: DEBUG nova.virt.hardware [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 639.807123] env[61868]: DEBUG nova.virt.hardware [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 639.807123] env[61868]: DEBUG nova.virt.hardware [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 639.807123] env[61868]: DEBUG nova.virt.hardware [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 639.807355] env[61868]: DEBUG nova.virt.hardware [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 639.807355] env[61868]: DEBUG nova.virt.hardware [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 639.807355] env[61868]: DEBUG nova.virt.hardware [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 639.807355] env[61868]: DEBUG nova.virt.hardware [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 639.807355] env[61868]: DEBUG nova.virt.hardware [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 639.808507] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abe9ca83-cb72-4e3c-b639-bea03124df2b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.818509] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-628a4931-9d72-4167-baf4-e14cd7d1b1f1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.876404] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.026177] env[61868]: DEBUG oslo_concurrency.lockutils [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.313508] env[61868]: DEBUG nova.network.neutron [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 640.458043] env[61868]: DEBUG nova.network.neutron [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.880144] env[61868]: DEBUG nova.compute.manager [req-e58e5f65-ce44-4a54-ba33-263669a67f3f req-6450f574-d720-4ed9-af65-f974bac135da service nova] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Received event network-changed-a4c26f3e-a6e6-45b9-b449-bf07b77ca99b {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 640.880144] env[61868]: DEBUG nova.compute.manager [req-e58e5f65-ce44-4a54-ba33-263669a67f3f req-6450f574-d720-4ed9-af65-f974bac135da service nova] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Refreshing instance network info cache due to event network-changed-a4c26f3e-a6e6-45b9-b449-bf07b77ca99b. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 640.880144] env[61868]: DEBUG oslo_concurrency.lockutils [req-e58e5f65-ce44-4a54-ba33-263669a67f3f req-6450f574-d720-4ed9-af65-f974bac135da service nova] Acquiring lock "refresh_cache-c46a8746-b4b2-46f8-b16a-a7ceb5ff4611" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 640.880144] env[61868]: DEBUG oslo_concurrency.lockutils [req-e58e5f65-ce44-4a54-ba33-263669a67f3f req-6450f574-d720-4ed9-af65-f974bac135da service nova] Acquired lock "refresh_cache-c46a8746-b4b2-46f8-b16a-a7ceb5ff4611" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.880144] env[61868]: DEBUG nova.network.neutron [req-e58e5f65-ce44-4a54-ba33-263669a67f3f req-6450f574-d720-4ed9-af65-f974bac135da service nova] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Refreshing network info cache for port a4c26f3e-a6e6-45b9-b449-bf07b77ca99b {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 640.961465] env[61868]: DEBUG oslo_concurrency.lockutils [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Releasing lock "refresh_cache-8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 640.961920] env[61868]: DEBUG nova.compute.manager [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 640.962086] env[61868]: DEBUG nova.compute.manager [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 640.962284] env[61868]: DEBUG nova.network.neutron [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 640.999651] env[61868]: DEBUG nova.network.neutron [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 641.290428] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb2006e1-eeed-4e6c-b21d-87caed70bd9a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.298393] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-202daf80-0ec7-4644-b39e-3f79ffc63a8c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.331898] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3ea8af2-f85b-4a95-930e-a9428ac7d849 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.340614] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8894ffd7-a127-47a2-bbd6-6ad2e6032aea {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.356648] env[61868]: DEBUG nova.compute.provider_tree [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 641.358193] env[61868]: DEBUG oslo_concurrency.lockutils [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Acquiring lock "7651f6d2-1825-46a1-9380-732b50abd3cc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.358448] env[61868]: DEBUG oslo_concurrency.lockutils [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Lock "7651f6d2-1825-46a1-9380-732b50abd3cc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 641.358649] env[61868]: DEBUG oslo_concurrency.lockutils [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Acquiring lock "7651f6d2-1825-46a1-9380-732b50abd3cc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.358828] env[61868]: DEBUG oslo_concurrency.lockutils [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Lock "7651f6d2-1825-46a1-9380-732b50abd3cc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 641.360235] env[61868]: DEBUG oslo_concurrency.lockutils [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Lock "7651f6d2-1825-46a1-9380-732b50abd3cc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 641.360953] env[61868]: INFO nova.compute.manager [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Terminating instance [ 641.362557] env[61868]: DEBUG oslo_concurrency.lockutils [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Acquiring lock "refresh_cache-7651f6d2-1825-46a1-9380-732b50abd3cc" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.362757] env[61868]: DEBUG oslo_concurrency.lockutils [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Acquired lock "refresh_cache-7651f6d2-1825-46a1-9380-732b50abd3cc" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.362864] env[61868]: DEBUG nova.network.neutron [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 642.041887] env[61868]: ERROR nova.compute.manager [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a4c26f3e-a6e6-45b9-b449-bf07b77ca99b, please check neutron logs for more information. [ 642.041887] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 642.041887] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 642.041887] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 642.041887] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 642.041887] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 642.041887] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 642.041887] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 642.041887] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 642.041887] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 642.041887] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 642.041887] env[61868]: ERROR nova.compute.manager raise self.value [ 642.041887] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 642.041887] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 642.041887] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 642.041887] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 642.042781] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 642.042781] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 642.042781] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a4c26f3e-a6e6-45b9-b449-bf07b77ca99b, please check neutron logs for more information. [ 642.042781] env[61868]: ERROR nova.compute.manager [ 642.042781] env[61868]: Traceback (most recent call last): [ 642.042781] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 642.042781] env[61868]: listener.cb(fileno) [ 642.042781] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 642.042781] env[61868]: result = function(*args, **kwargs) [ 642.042781] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 642.042781] env[61868]: return func(*args, **kwargs) [ 642.042781] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 642.042781] env[61868]: raise e [ 642.042781] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 642.042781] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 642.042781] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 642.042781] env[61868]: created_port_ids = self._update_ports_for_instance( [ 642.042781] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 642.042781] env[61868]: with excutils.save_and_reraise_exception(): [ 642.042781] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 642.042781] env[61868]: self.force_reraise() [ 642.042781] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 642.042781] env[61868]: raise self.value [ 642.042781] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 642.042781] env[61868]: updated_port = self._update_port( [ 642.042781] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 642.042781] env[61868]: _ensure_no_port_binding_failure(port) [ 642.042781] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 642.042781] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 642.043566] env[61868]: nova.exception.PortBindingFailed: Binding failed for port a4c26f3e-a6e6-45b9-b449-bf07b77ca99b, please check neutron logs for more information. [ 642.043566] env[61868]: Removing descriptor: 16 [ 642.046594] env[61868]: DEBUG nova.network.neutron [req-e58e5f65-ce44-4a54-ba33-263669a67f3f req-6450f574-d720-4ed9-af65-f974bac135da service nova] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.049299] env[61868]: DEBUG nova.network.neutron [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.051856] env[61868]: DEBUG nova.scheduler.client.report [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 642.057583] env[61868]: ERROR nova.compute.manager [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a4c26f3e-a6e6-45b9-b449-bf07b77ca99b, please check neutron logs for more information. [ 642.057583] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Traceback (most recent call last): [ 642.057583] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 642.057583] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] yield resources [ 642.057583] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 642.057583] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] self.driver.spawn(context, instance, image_meta, [ 642.057583] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 642.057583] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] self._vmops.spawn(context, instance, image_meta, injected_files, [ 642.057583] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 642.057583] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] vm_ref = self.build_virtual_machine(instance, [ 642.057583] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 642.057877] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] vif_infos = vmwarevif.get_vif_info(self._session, [ 642.057877] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 642.057877] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] for vif in network_info: [ 642.057877] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 642.057877] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] return self._sync_wrapper(fn, *args, **kwargs) [ 642.057877] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 642.057877] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] self.wait() [ 642.057877] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 642.057877] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] self[:] = self._gt.wait() [ 642.057877] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 642.057877] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] return self._exit_event.wait() [ 642.057877] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 642.057877] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] result = hub.switch() [ 642.058225] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 642.058225] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] return self.greenlet.switch() [ 642.058225] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 642.058225] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] result = function(*args, **kwargs) [ 642.058225] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 642.058225] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] return func(*args, **kwargs) [ 642.058225] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 642.058225] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] raise e [ 642.058225] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 642.058225] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] nwinfo = self.network_api.allocate_for_instance( [ 642.058225] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 642.058225] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] created_port_ids = self._update_ports_for_instance( [ 642.058225] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 642.058543] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] with excutils.save_and_reraise_exception(): [ 642.058543] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 642.058543] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] self.force_reraise() [ 642.058543] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 642.058543] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] raise self.value [ 642.058543] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 642.058543] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] updated_port = self._update_port( [ 642.058543] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 642.058543] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] _ensure_no_port_binding_failure(port) [ 642.058543] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 642.058543] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] raise exception.PortBindingFailed(port_id=port['id']) [ 642.058543] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] nova.exception.PortBindingFailed: Binding failed for port a4c26f3e-a6e6-45b9-b449-bf07b77ca99b, please check neutron logs for more information. [ 642.058543] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] [ 642.058915] env[61868]: INFO nova.compute.manager [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Terminating instance [ 642.058915] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] Acquiring lock "refresh_cache-c46a8746-b4b2-46f8-b16a-a7ceb5ff4611" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 642.112760] env[61868]: DEBUG nova.network.neutron [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.233090] env[61868]: DEBUG nova.network.neutron [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.256628] env[61868]: DEBUG nova.network.neutron [req-e58e5f65-ce44-4a54-ba33-263669a67f3f req-6450f574-d720-4ed9-af65-f974bac135da service nova] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.559597] env[61868]: INFO nova.compute.manager [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431] Took 1.60 seconds to deallocate network for instance. [ 642.563126] env[61868]: DEBUG oslo_concurrency.lockutils [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.803s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.563649] env[61868]: DEBUG nova.compute.manager [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 642.566324] env[61868]: DEBUG oslo_concurrency.lockutils [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.114s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.735961] env[61868]: DEBUG oslo_concurrency.lockutils [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Releasing lock "refresh_cache-7651f6d2-1825-46a1-9380-732b50abd3cc" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 642.737434] env[61868]: DEBUG nova.compute.manager [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 642.737615] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 642.738840] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0dd296f-36a1-4cee-a627-bc2a2734b4d5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.746767] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 642.747050] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1a591a37-5167-4d88-b57d-fbeefba181af {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.753623] env[61868]: DEBUG oslo_vmware.api [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Waiting for the task: (returnval){ [ 642.753623] env[61868]: value = "task-1315305" [ 642.753623] env[61868]: _type = "Task" [ 642.753623] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.761706] env[61868]: DEBUG oslo_vmware.api [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315305, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.764450] env[61868]: DEBUG oslo_concurrency.lockutils [req-e58e5f65-ce44-4a54-ba33-263669a67f3f req-6450f574-d720-4ed9-af65-f974bac135da service nova] Releasing lock "refresh_cache-c46a8746-b4b2-46f8-b16a-a7ceb5ff4611" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 642.764450] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] Acquired lock "refresh_cache-c46a8746-b4b2-46f8-b16a-a7ceb5ff4611" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.764450] env[61868]: DEBUG nova.network.neutron [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 643.070717] env[61868]: DEBUG nova.compute.utils [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 643.072966] env[61868]: DEBUG nova.compute.manager [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 643.072966] env[61868]: DEBUG nova.network.neutron [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 643.103168] env[61868]: DEBUG nova.compute.manager [req-b85f7907-64ab-464e-86c2-8d9a28785e82 req-94fea648-215d-4357-a4fe-8c77780310fc service nova] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Received event network-vif-deleted-a4c26f3e-a6e6-45b9-b449-bf07b77ca99b {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 643.156113] env[61868]: DEBUG nova.policy [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e5ea0c10f8de45db92d7f106d3544a44', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '500e93f8728341bd83158cac3d8b03e4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 643.271615] env[61868]: DEBUG oslo_vmware.api [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315305, 'name': PowerOffVM_Task, 'duration_secs': 0.123589} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.271615] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 643.271615] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 643.271615] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9c29dd8b-34f2-4822-9db9-3f102568b0a2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.288641] env[61868]: DEBUG nova.network.neutron [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 643.300031] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 643.300497] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Deleting contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 643.300581] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Deleting the datastore file [datastore2] 7651f6d2-1825-46a1-9380-732b50abd3cc {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 643.300862] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9f7cd5fd-c795-4512-a27a-e415f9c6e194 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.309473] env[61868]: DEBUG oslo_vmware.api [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Waiting for the task: (returnval){ [ 643.309473] env[61868]: value = "task-1315307" [ 643.309473] env[61868]: _type = "Task" [ 643.309473] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.324037] env[61868]: DEBUG oslo_vmware.api [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315307, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.416045] env[61868]: DEBUG nova.network.neutron [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.579166] env[61868]: DEBUG nova.compute.manager [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 643.586275] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5763194e-55a3-494c-b3cf-fbbfcdbc137a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.595485] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6f5421f-a7c8-448f-bbbf-7a457084fa91 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.633665] env[61868]: INFO nova.scheduler.client.report [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Deleted allocations for instance 8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431 [ 643.639648] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70054581-67fc-47ea-99d7-f4e6fde39628 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.651445] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d4780fb-f115-4a7d-8c78-69efb409284c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.655298] env[61868]: DEBUG nova.network.neutron [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Successfully created port: 6cdafb9f-87e5-4c0a-8bd1-1fcec1a98833 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 643.669627] env[61868]: DEBUG nova.compute.provider_tree [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 643.718926] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Acquiring lock "29530178-d69c-4aed-9061-d3d1cfa954d0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.719379] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Lock "29530178-d69c-4aed-9061-d3d1cfa954d0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 643.825255] env[61868]: DEBUG oslo_vmware.api [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Task: {'id': task-1315307, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.110368} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.825771] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 643.825771] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Deleted contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 643.825900] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 643.825986] env[61868]: INFO nova.compute.manager [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Took 1.09 seconds to destroy the instance on the hypervisor. [ 643.826233] env[61868]: DEBUG oslo.service.loopingcall [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 643.826410] env[61868]: DEBUG nova.compute.manager [-] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 643.826606] env[61868]: DEBUG nova.network.neutron [-] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 643.854277] env[61868]: DEBUG nova.network.neutron [-] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 643.918787] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] Releasing lock "refresh_cache-c46a8746-b4b2-46f8-b16a-a7ceb5ff4611" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.919243] env[61868]: DEBUG nova.compute.manager [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 643.920026] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 643.920026] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-810f981b-5cfd-4a7d-8cc9-d9e9de012f59 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.928659] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2d4bc93-79ef-4069-8087-5a51a64ab300 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.951370] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c46a8746-b4b2-46f8-b16a-a7ceb5ff4611 could not be found. [ 643.951747] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 643.952066] env[61868]: INFO nova.compute.manager [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Took 0.03 seconds to destroy the instance on the hypervisor. [ 643.952462] env[61868]: DEBUG oslo.service.loopingcall [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 643.952840] env[61868]: DEBUG nova.compute.manager [-] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 643.952967] env[61868]: DEBUG nova.network.neutron [-] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 643.982608] env[61868]: DEBUG nova.network.neutron [-] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 644.146945] env[61868]: DEBUG oslo_concurrency.lockutils [None req-96f55f48-a6f2-4ca6-beb6-d7fec941c921 tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Lock "8b9c4ecf-e2a3-42bf-ae63-3ec4bc637431" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 81.735s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 644.174338] env[61868]: DEBUG nova.scheduler.client.report [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 644.357423] env[61868]: DEBUG nova.network.neutron [-] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.392228] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 644.392488] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 644.485345] env[61868]: DEBUG nova.network.neutron [-] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.594022] env[61868]: DEBUG nova.compute.manager [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 644.623475] env[61868]: DEBUG nova.virt.hardware [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 644.623880] env[61868]: DEBUG nova.virt.hardware [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 644.624092] env[61868]: DEBUG nova.virt.hardware [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 644.624939] env[61868]: DEBUG nova.virt.hardware [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 644.625137] env[61868]: DEBUG nova.virt.hardware [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 644.625299] env[61868]: DEBUG nova.virt.hardware [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 644.625944] env[61868]: DEBUG nova.virt.hardware [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 644.625944] env[61868]: DEBUG nova.virt.hardware [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 644.626105] env[61868]: DEBUG nova.virt.hardware [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 644.626181] env[61868]: DEBUG nova.virt.hardware [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 644.626350] env[61868]: DEBUG nova.virt.hardware [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 644.629273] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4361e7b3-5ba5-4456-998b-4ff552570e42 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.637884] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cf0a55c-2e54-4926-910d-9ce7c02988c8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.657828] env[61868]: DEBUG nova.compute.manager [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 644.680917] env[61868]: DEBUG oslo_concurrency.lockutils [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.114s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 644.683019] env[61868]: ERROR nova.compute.manager [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d9e16fa5-610c-4053-8a20-1c33bf8d6324, please check neutron logs for more information. [ 644.683019] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Traceback (most recent call last): [ 644.683019] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 644.683019] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] self.driver.spawn(context, instance, image_meta, [ 644.683019] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 644.683019] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 644.683019] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 644.683019] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] vm_ref = self.build_virtual_machine(instance, [ 644.683019] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 644.683019] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] vif_infos = vmwarevif.get_vif_info(self._session, [ 644.683019] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 644.683335] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] for vif in network_info: [ 644.683335] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 644.683335] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] return self._sync_wrapper(fn, *args, **kwargs) [ 644.683335] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 644.683335] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] self.wait() [ 644.683335] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 644.683335] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] self[:] = self._gt.wait() [ 644.683335] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 644.683335] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] return self._exit_event.wait() [ 644.683335] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 644.683335] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] result = hub.switch() [ 644.683335] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 644.683335] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] return self.greenlet.switch() [ 644.683628] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 644.683628] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] result = function(*args, **kwargs) [ 644.683628] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 644.683628] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] return func(*args, **kwargs) [ 644.683628] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 644.683628] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] raise e [ 644.683628] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 644.683628] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] nwinfo = self.network_api.allocate_for_instance( [ 644.683628] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 644.683628] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] created_port_ids = self._update_ports_for_instance( [ 644.683628] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 644.683628] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] with excutils.save_and_reraise_exception(): [ 644.683628] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 644.683922] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] self.force_reraise() [ 644.683922] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 644.683922] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] raise self.value [ 644.683922] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 644.683922] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] updated_port = self._update_port( [ 644.683922] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 644.683922] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] _ensure_no_port_binding_failure(port) [ 644.683922] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 644.683922] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] raise exception.PortBindingFailed(port_id=port['id']) [ 644.683922] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] nova.exception.PortBindingFailed: Binding failed for port d9e16fa5-610c-4053-8a20-1c33bf8d6324, please check neutron logs for more information. [ 644.683922] env[61868]: ERROR nova.compute.manager [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] [ 644.684181] env[61868]: DEBUG nova.compute.utils [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Binding failed for port d9e16fa5-610c-4053-8a20-1c33bf8d6324, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 644.684181] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.709s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 644.685615] env[61868]: INFO nova.compute.claims [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 644.687811] env[61868]: DEBUG nova.compute.manager [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Build of instance 7c2937c6-e6b7-4716-80fa-5538641bc8dc was re-scheduled: Binding failed for port d9e16fa5-610c-4053-8a20-1c33bf8d6324, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 644.688489] env[61868]: DEBUG nova.compute.manager [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 644.688489] env[61868]: DEBUG oslo_concurrency.lockutils [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Acquiring lock "refresh_cache-7c2937c6-e6b7-4716-80fa-5538641bc8dc" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 644.688599] env[61868]: DEBUG oslo_concurrency.lockutils [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Acquired lock "refresh_cache-7c2937c6-e6b7-4716-80fa-5538641bc8dc" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 644.688661] env[61868]: DEBUG nova.network.neutron [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 644.861170] env[61868]: INFO nova.compute.manager [-] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Took 1.03 seconds to deallocate network for instance. [ 644.904367] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 644.904574] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Starting heal instance info cache {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 644.904698] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Rebuilding the list of instances to heal {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 644.988431] env[61868]: INFO nova.compute.manager [-] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Took 1.04 seconds to deallocate network for instance. [ 644.990818] env[61868]: DEBUG nova.compute.claims [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 644.991021] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.126754] env[61868]: DEBUG nova.compute.manager [req-e3e1a43c-80ed-44e8-a42b-79e6d6dc6af5 req-0fab5bb6-581a-4cbe-a709-32b403131d56 service nova] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Received event network-changed-6cdafb9f-87e5-4c0a-8bd1-1fcec1a98833 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 645.126948] env[61868]: DEBUG nova.compute.manager [req-e3e1a43c-80ed-44e8-a42b-79e6d6dc6af5 req-0fab5bb6-581a-4cbe-a709-32b403131d56 service nova] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Refreshing instance network info cache due to event network-changed-6cdafb9f-87e5-4c0a-8bd1-1fcec1a98833. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 645.127365] env[61868]: DEBUG oslo_concurrency.lockutils [req-e3e1a43c-80ed-44e8-a42b-79e6d6dc6af5 req-0fab5bb6-581a-4cbe-a709-32b403131d56 service nova] Acquiring lock "refresh_cache-1586080c-3406-47f0-bbb9-0c50082eaf8d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 645.127517] env[61868]: DEBUG oslo_concurrency.lockutils [req-e3e1a43c-80ed-44e8-a42b-79e6d6dc6af5 req-0fab5bb6-581a-4cbe-a709-32b403131d56 service nova] Acquired lock "refresh_cache-1586080c-3406-47f0-bbb9-0c50082eaf8d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.127680] env[61868]: DEBUG nova.network.neutron [req-e3e1a43c-80ed-44e8-a42b-79e6d6dc6af5 req-0fab5bb6-581a-4cbe-a709-32b403131d56 service nova] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Refreshing network info cache for port 6cdafb9f-87e5-4c0a-8bd1-1fcec1a98833 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 645.187775] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.232974] env[61868]: DEBUG nova.network.neutron [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 645.347188] env[61868]: DEBUG nova.network.neutron [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.367724] env[61868]: DEBUG oslo_concurrency.lockutils [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.397044] env[61868]: ERROR nova.compute.manager [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6cdafb9f-87e5-4c0a-8bd1-1fcec1a98833, please check neutron logs for more information. [ 645.397044] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 645.397044] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 645.397044] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 645.397044] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 645.397044] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 645.397044] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 645.397044] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 645.397044] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 645.397044] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 645.397044] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 645.397044] env[61868]: ERROR nova.compute.manager raise self.value [ 645.397044] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 645.397044] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 645.397044] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 645.397044] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 645.397493] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 645.397493] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 645.397493] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6cdafb9f-87e5-4c0a-8bd1-1fcec1a98833, please check neutron logs for more information. [ 645.397493] env[61868]: ERROR nova.compute.manager [ 645.397493] env[61868]: Traceback (most recent call last): [ 645.397493] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 645.397493] env[61868]: listener.cb(fileno) [ 645.397493] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 645.397493] env[61868]: result = function(*args, **kwargs) [ 645.397493] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 645.397493] env[61868]: return func(*args, **kwargs) [ 645.397493] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 645.397493] env[61868]: raise e [ 645.397493] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 645.397493] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 645.397493] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 645.397493] env[61868]: created_port_ids = self._update_ports_for_instance( [ 645.397493] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 645.397493] env[61868]: with excutils.save_and_reraise_exception(): [ 645.397493] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 645.397493] env[61868]: self.force_reraise() [ 645.397493] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 645.397493] env[61868]: raise self.value [ 645.397493] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 645.397493] env[61868]: updated_port = self._update_port( [ 645.397493] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 645.397493] env[61868]: _ensure_no_port_binding_failure(port) [ 645.397493] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 645.397493] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 645.398312] env[61868]: nova.exception.PortBindingFailed: Binding failed for port 6cdafb9f-87e5-4c0a-8bd1-1fcec1a98833, please check neutron logs for more information. [ 645.398312] env[61868]: Removing descriptor: 15 [ 645.398312] env[61868]: ERROR nova.compute.manager [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6cdafb9f-87e5-4c0a-8bd1-1fcec1a98833, please check neutron logs for more information. [ 645.398312] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Traceback (most recent call last): [ 645.398312] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 645.398312] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] yield resources [ 645.398312] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 645.398312] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] self.driver.spawn(context, instance, image_meta, [ 645.398312] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 645.398312] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 645.398312] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 645.398312] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] vm_ref = self.build_virtual_machine(instance, [ 645.398638] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 645.398638] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] vif_infos = vmwarevif.get_vif_info(self._session, [ 645.398638] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 645.398638] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] for vif in network_info: [ 645.398638] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 645.398638] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] return self._sync_wrapper(fn, *args, **kwargs) [ 645.398638] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 645.398638] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] self.wait() [ 645.398638] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 645.398638] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] self[:] = self._gt.wait() [ 645.398638] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 645.398638] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] return self._exit_event.wait() [ 645.398638] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 645.398980] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] result = hub.switch() [ 645.398980] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 645.398980] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] return self.greenlet.switch() [ 645.398980] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 645.398980] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] result = function(*args, **kwargs) [ 645.398980] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 645.398980] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] return func(*args, **kwargs) [ 645.398980] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 645.398980] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] raise e [ 645.398980] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 645.398980] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] nwinfo = self.network_api.allocate_for_instance( [ 645.398980] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 645.398980] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] created_port_ids = self._update_ports_for_instance( [ 645.399336] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 645.399336] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] with excutils.save_and_reraise_exception(): [ 645.399336] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 645.399336] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] self.force_reraise() [ 645.399336] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 645.399336] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] raise self.value [ 645.399336] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 645.399336] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] updated_port = self._update_port( [ 645.399336] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 645.399336] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] _ensure_no_port_binding_failure(port) [ 645.399336] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 645.399336] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] raise exception.PortBindingFailed(port_id=port['id']) [ 645.399651] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] nova.exception.PortBindingFailed: Binding failed for port 6cdafb9f-87e5-4c0a-8bd1-1fcec1a98833, please check neutron logs for more information. [ 645.399651] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] [ 645.399651] env[61868]: INFO nova.compute.manager [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Terminating instance [ 645.401049] env[61868]: DEBUG oslo_concurrency.lockutils [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] Acquiring lock "refresh_cache-1586080c-3406-47f0-bbb9-0c50082eaf8d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 645.411309] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Skipping network cache update for instance because it is Building. {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 645.412230] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Skipping network cache update for instance because it is Building. {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 645.412230] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Skipping network cache update for instance because it is Building. {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 645.412230] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Skipping network cache update for instance because it is Building. {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 645.443085] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "refresh_cache-7651f6d2-1825-46a1-9380-732b50abd3cc" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 645.443238] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquired lock "refresh_cache-7651f6d2-1825-46a1-9380-732b50abd3cc" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.443383] env[61868]: DEBUG nova.network.neutron [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Forcefully refreshing network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 645.443535] env[61868]: DEBUG nova.objects.instance [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lazy-loading 'info_cache' on Instance uuid 7651f6d2-1825-46a1-9380-732b50abd3cc {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 645.661664] env[61868]: DEBUG nova.network.neutron [req-e3e1a43c-80ed-44e8-a42b-79e6d6dc6af5 req-0fab5bb6-581a-4cbe-a709-32b403131d56 service nova] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 645.800562] env[61868]: DEBUG nova.network.neutron [req-e3e1a43c-80ed-44e8-a42b-79e6d6dc6af5 req-0fab5bb6-581a-4cbe-a709-32b403131d56 service nova] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.849931] env[61868]: DEBUG oslo_concurrency.lockutils [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Releasing lock "refresh_cache-7c2937c6-e6b7-4716-80fa-5538641bc8dc" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 645.850211] env[61868]: DEBUG nova.compute.manager [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 645.850421] env[61868]: DEBUG nova.compute.manager [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 645.850707] env[61868]: DEBUG nova.network.neutron [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 645.911166] env[61868]: DEBUG nova.network.neutron [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 646.212468] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8061f12c-ebd3-4445-ba0e-08e230465c5c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.220417] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c25acc37-3b1b-402a-9b94-82f614eefdc0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.255389] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8340bb1c-b9e2-4d84-a23e-99cb665c8d8e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.263953] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-377a7936-ce9e-4e8c-ac9d-53bda81b73eb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.277645] env[61868]: DEBUG nova.compute.provider_tree [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 646.305601] env[61868]: DEBUG oslo_concurrency.lockutils [req-e3e1a43c-80ed-44e8-a42b-79e6d6dc6af5 req-0fab5bb6-581a-4cbe-a709-32b403131d56 service nova] Releasing lock "refresh_cache-1586080c-3406-47f0-bbb9-0c50082eaf8d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 646.306008] env[61868]: DEBUG oslo_concurrency.lockutils [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] Acquired lock "refresh_cache-1586080c-3406-47f0-bbb9-0c50082eaf8d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.306206] env[61868]: DEBUG nova.network.neutron [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 646.413581] env[61868]: DEBUG nova.network.neutron [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.473381] env[61868]: DEBUG nova.network.neutron [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 646.782082] env[61868]: DEBUG nova.scheduler.client.report [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 646.827258] env[61868]: DEBUG nova.network.neutron [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 646.916483] env[61868]: INFO nova.compute.manager [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] [instance: 7c2937c6-e6b7-4716-80fa-5538641bc8dc] Took 1.07 seconds to deallocate network for instance. [ 646.954204] env[61868]: DEBUG nova.network.neutron [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.057486] env[61868]: DEBUG nova.network.neutron [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.151190] env[61868]: DEBUG nova.compute.manager [req-554cb3e0-befe-4dab-91de-d3b8ff6d02f4 req-cc9c3ed5-516e-478b-a828-07dc78c5ce9e service nova] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Received event network-vif-deleted-6cdafb9f-87e5-4c0a-8bd1-1fcec1a98833 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 647.287299] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.603s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.287786] env[61868]: DEBUG nova.compute.manager [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 647.294318] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 16.096s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.294651] env[61868]: DEBUG nova.objects.instance [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61868) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 647.457596] env[61868]: DEBUG oslo_concurrency.lockutils [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] Releasing lock "refresh_cache-1586080c-3406-47f0-bbb9-0c50082eaf8d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 647.458120] env[61868]: DEBUG nova.compute.manager [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 647.458331] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 647.458644] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-33edca0f-8141-43ff-a7ae-0ca9dcc26043 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.469762] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6afc51ed-b210-41a2-847f-ce827f3b073d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.496821] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1586080c-3406-47f0-bbb9-0c50082eaf8d could not be found. [ 647.497064] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 647.497245] env[61868]: INFO nova.compute.manager [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Took 0.04 seconds to destroy the instance on the hypervisor. [ 647.497482] env[61868]: DEBUG oslo.service.loopingcall [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 647.497692] env[61868]: DEBUG nova.compute.manager [-] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 647.497782] env[61868]: DEBUG nova.network.neutron [-] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 647.522839] env[61868]: DEBUG nova.network.neutron [-] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 647.563258] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Releasing lock "refresh_cache-7651f6d2-1825-46a1-9380-732b50abd3cc" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 647.563258] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Updated the network info_cache for instance {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 647.563258] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 647.563258] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 647.563258] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 647.563258] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 647.563576] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 647.563576] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 647.563576] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61868) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 647.563576] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager.update_available_resource {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 647.799904] env[61868]: DEBUG nova.compute.utils [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 647.806199] env[61868]: DEBUG nova.compute.manager [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 647.810667] env[61868]: DEBUG nova.network.neutron [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 647.883080] env[61868]: DEBUG nova.policy [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1f4a142657d44ea2ba4f37d07dafc653', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cf8e0232b4cb44dbb966c66c46405c78', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 647.958835] env[61868]: INFO nova.scheduler.client.report [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Deleted allocations for instance 7c2937c6-e6b7-4716-80fa-5538641bc8dc [ 648.025853] env[61868]: DEBUG nova.network.neutron [-] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.069522] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 648.306090] env[61868]: DEBUG nova.compute.manager [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 648.318071] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ceace5f5-a5b2-4252-9490-a22844bfb808 tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.021s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 648.318071] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.426s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 648.390285] env[61868]: DEBUG nova.network.neutron [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Successfully created port: 0aeb4b03-3fbd-4eb0-9261-fb6dc4b6077e {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 648.438679] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] Acquiring lock "e023cfb8-5f1c-41d9-9e43-d10e073234ea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 648.438765] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] Lock "e023cfb8-5f1c-41d9-9e43-d10e073234ea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 648.474334] env[61868]: DEBUG oslo_concurrency.lockutils [None req-50e9d864-702d-4b97-a0eb-d9127506afbc tempest-ListImageFiltersTestJSON-1332600325 tempest-ListImageFiltersTestJSON-1332600325-project-member] Lock "7c2937c6-e6b7-4716-80fa-5538641bc8dc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 84.453s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 648.528491] env[61868]: INFO nova.compute.manager [-] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Took 1.03 seconds to deallocate network for instance. [ 648.531379] env[61868]: DEBUG nova.compute.claims [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 648.531634] env[61868]: DEBUG oslo_concurrency.lockutils [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 648.978067] env[61868]: DEBUG nova.compute.manager [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 649.318170] env[61868]: DEBUG nova.compute.manager [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 649.344506] env[61868]: DEBUG nova.virt.hardware [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 649.344637] env[61868]: DEBUG nova.virt.hardware [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 649.344706] env[61868]: DEBUG nova.virt.hardware [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 649.344907] env[61868]: DEBUG nova.virt.hardware [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 649.345184] env[61868]: DEBUG nova.virt.hardware [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 649.345342] env[61868]: DEBUG nova.virt.hardware [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 649.345559] env[61868]: DEBUG nova.virt.hardware [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 649.345723] env[61868]: DEBUG nova.virt.hardware [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 649.345904] env[61868]: DEBUG nova.virt.hardware [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 649.346092] env[61868]: DEBUG nova.virt.hardware [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 649.346282] env[61868]: DEBUG nova.virt.hardware [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 649.347224] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fede555e-bb55-41e6-a7d1-d63a20611cf7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.359931] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07e05913-39f9-4f48-9f7d-d066fee1d07c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.380715] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a26c06ee-d2ef-4f43-b56e-b8838e6099ea {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.389066] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eeca4bb-9236-4c08-8647-6c89c66bfa51 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.419489] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fe89575-0579-4ca2-8f25-73524aabf137 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.428231] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db772f4c-d86f-44b9-aee5-ed48d11cb901 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.440653] env[61868]: DEBUG nova.compute.provider_tree [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 649.460075] env[61868]: DEBUG nova.compute.manager [req-8e43fd95-04ce-4749-9a50-33a77b710c1c req-23b49c83-d657-408a-b86c-6f5927d3df2e service nova] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Received event network-changed-0aeb4b03-3fbd-4eb0-9261-fb6dc4b6077e {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 649.460303] env[61868]: DEBUG nova.compute.manager [req-8e43fd95-04ce-4749-9a50-33a77b710c1c req-23b49c83-d657-408a-b86c-6f5927d3df2e service nova] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Refreshing instance network info cache due to event network-changed-0aeb4b03-3fbd-4eb0-9261-fb6dc4b6077e. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 649.460531] env[61868]: DEBUG oslo_concurrency.lockutils [req-8e43fd95-04ce-4749-9a50-33a77b710c1c req-23b49c83-d657-408a-b86c-6f5927d3df2e service nova] Acquiring lock "refresh_cache-ae5ec212-9dbe-4a11-94a3-8aa4576ca4db" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 649.460665] env[61868]: DEBUG oslo_concurrency.lockutils [req-8e43fd95-04ce-4749-9a50-33a77b710c1c req-23b49c83-d657-408a-b86c-6f5927d3df2e service nova] Acquired lock "refresh_cache-ae5ec212-9dbe-4a11-94a3-8aa4576ca4db" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 649.460819] env[61868]: DEBUG nova.network.neutron [req-8e43fd95-04ce-4749-9a50-33a77b710c1c req-23b49c83-d657-408a-b86c-6f5927d3df2e service nova] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Refreshing network info cache for port 0aeb4b03-3fbd-4eb0-9261-fb6dc4b6077e {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 649.504772] env[61868]: DEBUG oslo_concurrency.lockutils [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 649.516553] env[61868]: ERROR nova.compute.manager [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0aeb4b03-3fbd-4eb0-9261-fb6dc4b6077e, please check neutron logs for more information. [ 649.516553] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 649.516553] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 649.516553] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 649.516553] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 649.516553] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 649.516553] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 649.516553] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 649.516553] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 649.516553] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 649.516553] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 649.516553] env[61868]: ERROR nova.compute.manager raise self.value [ 649.516553] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 649.516553] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 649.516553] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 649.516553] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 649.517108] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 649.517108] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 649.517108] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0aeb4b03-3fbd-4eb0-9261-fb6dc4b6077e, please check neutron logs for more information. [ 649.517108] env[61868]: ERROR nova.compute.manager [ 649.517108] env[61868]: Traceback (most recent call last): [ 649.517108] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 649.517108] env[61868]: listener.cb(fileno) [ 649.517108] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 649.517108] env[61868]: result = function(*args, **kwargs) [ 649.517108] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 649.517108] env[61868]: return func(*args, **kwargs) [ 649.517108] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 649.517108] env[61868]: raise e [ 649.517108] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 649.517108] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 649.517108] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 649.517108] env[61868]: created_port_ids = self._update_ports_for_instance( [ 649.517108] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 649.517108] env[61868]: with excutils.save_and_reraise_exception(): [ 649.517108] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 649.517108] env[61868]: self.force_reraise() [ 649.517108] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 649.517108] env[61868]: raise self.value [ 649.517108] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 649.517108] env[61868]: updated_port = self._update_port( [ 649.517108] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 649.517108] env[61868]: _ensure_no_port_binding_failure(port) [ 649.517108] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 649.517108] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 649.517880] env[61868]: nova.exception.PortBindingFailed: Binding failed for port 0aeb4b03-3fbd-4eb0-9261-fb6dc4b6077e, please check neutron logs for more information. [ 649.517880] env[61868]: Removing descriptor: 15 [ 649.517880] env[61868]: ERROR nova.compute.manager [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0aeb4b03-3fbd-4eb0-9261-fb6dc4b6077e, please check neutron logs for more information. [ 649.517880] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Traceback (most recent call last): [ 649.517880] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 649.517880] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] yield resources [ 649.517880] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 649.517880] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] self.driver.spawn(context, instance, image_meta, [ 649.517880] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 649.517880] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] self._vmops.spawn(context, instance, image_meta, injected_files, [ 649.517880] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 649.517880] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] vm_ref = self.build_virtual_machine(instance, [ 649.518246] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 649.518246] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] vif_infos = vmwarevif.get_vif_info(self._session, [ 649.518246] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 649.518246] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] for vif in network_info: [ 649.518246] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 649.518246] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] return self._sync_wrapper(fn, *args, **kwargs) [ 649.518246] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 649.518246] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] self.wait() [ 649.518246] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 649.518246] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] self[:] = self._gt.wait() [ 649.518246] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 649.518246] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] return self._exit_event.wait() [ 649.518246] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 649.518609] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] result = hub.switch() [ 649.518609] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 649.518609] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] return self.greenlet.switch() [ 649.518609] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 649.518609] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] result = function(*args, **kwargs) [ 649.518609] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 649.518609] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] return func(*args, **kwargs) [ 649.518609] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 649.518609] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] raise e [ 649.518609] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 649.518609] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] nwinfo = self.network_api.allocate_for_instance( [ 649.518609] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 649.518609] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] created_port_ids = self._update_ports_for_instance( [ 649.518963] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 649.518963] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] with excutils.save_and_reraise_exception(): [ 649.518963] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 649.518963] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] self.force_reraise() [ 649.518963] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 649.518963] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] raise self.value [ 649.518963] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 649.518963] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] updated_port = self._update_port( [ 649.518963] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 649.518963] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] _ensure_no_port_binding_failure(port) [ 649.518963] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 649.518963] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] raise exception.PortBindingFailed(port_id=port['id']) [ 649.519352] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] nova.exception.PortBindingFailed: Binding failed for port 0aeb4b03-3fbd-4eb0-9261-fb6dc4b6077e, please check neutron logs for more information. [ 649.519352] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] [ 649.519352] env[61868]: INFO nova.compute.manager [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Terminating instance [ 649.522809] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] Acquiring lock "refresh_cache-ae5ec212-9dbe-4a11-94a3-8aa4576ca4db" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 649.944526] env[61868]: DEBUG nova.scheduler.client.report [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 649.985241] env[61868]: DEBUG nova.network.neutron [req-8e43fd95-04ce-4749-9a50-33a77b710c1c req-23b49c83-d657-408a-b86c-6f5927d3df2e service nova] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 650.069317] env[61868]: DEBUG nova.network.neutron [req-8e43fd95-04ce-4749-9a50-33a77b710c1c req-23b49c83-d657-408a-b86c-6f5927d3df2e service nova] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.451512] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.135s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 650.452149] env[61868]: ERROR nova.compute.manager [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 393c5ca8-99b5-438b-b7d1-7a6c503c8b5a, please check neutron logs for more information. [ 650.452149] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Traceback (most recent call last): [ 650.452149] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 650.452149] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] self.driver.spawn(context, instance, image_meta, [ 650.452149] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 650.452149] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] self._vmops.spawn(context, instance, image_meta, injected_files, [ 650.452149] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 650.452149] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] vm_ref = self.build_virtual_machine(instance, [ 650.452149] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 650.452149] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] vif_infos = vmwarevif.get_vif_info(self._session, [ 650.452149] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 650.452446] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] for vif in network_info: [ 650.452446] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 650.452446] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] return self._sync_wrapper(fn, *args, **kwargs) [ 650.452446] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 650.452446] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] self.wait() [ 650.452446] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 650.452446] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] self[:] = self._gt.wait() [ 650.452446] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 650.452446] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] return self._exit_event.wait() [ 650.452446] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 650.452446] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] result = hub.switch() [ 650.452446] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 650.452446] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] return self.greenlet.switch() [ 650.452743] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 650.452743] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] result = function(*args, **kwargs) [ 650.452743] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 650.452743] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] return func(*args, **kwargs) [ 650.452743] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 650.452743] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] raise e [ 650.452743] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 650.452743] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] nwinfo = self.network_api.allocate_for_instance( [ 650.452743] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 650.452743] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] created_port_ids = self._update_ports_for_instance( [ 650.452743] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 650.452743] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] with excutils.save_and_reraise_exception(): [ 650.452743] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 650.453081] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] self.force_reraise() [ 650.453081] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 650.453081] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] raise self.value [ 650.453081] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 650.453081] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] updated_port = self._update_port( [ 650.453081] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 650.453081] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] _ensure_no_port_binding_failure(port) [ 650.453081] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 650.453081] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] raise exception.PortBindingFailed(port_id=port['id']) [ 650.453081] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] nova.exception.PortBindingFailed: Binding failed for port 393c5ca8-99b5-438b-b7d1-7a6c503c8b5a, please check neutron logs for more information. [ 650.453081] env[61868]: ERROR nova.compute.manager [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] [ 650.453360] env[61868]: DEBUG nova.compute.utils [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Binding failed for port 393c5ca8-99b5-438b-b7d1-7a6c503c8b5a, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 650.454928] env[61868]: DEBUG oslo_concurrency.lockutils [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.807s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 650.457676] env[61868]: INFO nova.compute.claims [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 650.462049] env[61868]: DEBUG nova.compute.manager [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Build of instance 4c98c21e-2918-4b55-b2b5-db14630a36ef was re-scheduled: Binding failed for port 393c5ca8-99b5-438b-b7d1-7a6c503c8b5a, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 650.462616] env[61868]: DEBUG nova.compute.manager [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 650.463285] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] Acquiring lock "refresh_cache-4c98c21e-2918-4b55-b2b5-db14630a36ef" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 650.463507] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] Acquired lock "refresh_cache-4c98c21e-2918-4b55-b2b5-db14630a36ef" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.463705] env[61868]: DEBUG nova.network.neutron [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 650.573741] env[61868]: DEBUG oslo_concurrency.lockutils [req-8e43fd95-04ce-4749-9a50-33a77b710c1c req-23b49c83-d657-408a-b86c-6f5927d3df2e service nova] Releasing lock "refresh_cache-ae5ec212-9dbe-4a11-94a3-8aa4576ca4db" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 650.573741] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] Acquired lock "refresh_cache-ae5ec212-9dbe-4a11-94a3-8aa4576ca4db" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.573741] env[61868]: DEBUG nova.network.neutron [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 650.991051] env[61868]: DEBUG nova.network.neutron [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 651.080174] env[61868]: DEBUG nova.network.neutron [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.100534] env[61868]: DEBUG nova.network.neutron [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 651.183081] env[61868]: DEBUG nova.network.neutron [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.508146] env[61868]: DEBUG nova.compute.manager [req-27b1d258-5737-4d49-9587-f1f0a87885af req-c3d81966-dfcf-4b9c-8576-db6beb77bd39 service nova] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Received event network-vif-deleted-0aeb4b03-3fbd-4eb0-9261-fb6dc4b6077e {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 651.588123] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] Releasing lock "refresh_cache-4c98c21e-2918-4b55-b2b5-db14630a36ef" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 651.588290] env[61868]: DEBUG nova.compute.manager [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 651.588494] env[61868]: DEBUG nova.compute.manager [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 651.588665] env[61868]: DEBUG nova.network.neutron [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 651.607279] env[61868]: DEBUG nova.network.neutron [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 651.687891] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] Releasing lock "refresh_cache-ae5ec212-9dbe-4a11-94a3-8aa4576ca4db" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 651.687891] env[61868]: DEBUG nova.compute.manager [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 651.687891] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 651.688594] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cfc8dd1b-8fd3-4e62-8389-d2bf7953b8aa {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.698024] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a55eaf5-7694-4cee-92c5-df6230881c66 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.724600] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ae5ec212-9dbe-4a11-94a3-8aa4576ca4db could not be found. [ 651.724951] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 651.725186] env[61868]: INFO nova.compute.manager [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Took 0.04 seconds to destroy the instance on the hypervisor. [ 651.725450] env[61868]: DEBUG oslo.service.loopingcall [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 651.725694] env[61868]: DEBUG nova.compute.manager [-] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 651.725915] env[61868]: DEBUG nova.network.neutron [-] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 651.749933] env[61868]: DEBUG nova.network.neutron [-] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 651.971902] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4eb7814-cec2-46e7-b2e5-6656f5640f4e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.979630] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6ea0373-7905-4f68-9ba5-1cf9e769b740 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.012839] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fa0cd1d-cd74-4038-997e-33ee787aa050 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.020878] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba825d0e-0723-4610-837b-fdcbe80e6c11 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.034939] env[61868]: DEBUG nova.compute.provider_tree [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 652.109920] env[61868]: DEBUG nova.network.neutron [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.253417] env[61868]: DEBUG nova.network.neutron [-] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.538562] env[61868]: DEBUG nova.scheduler.client.report [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 652.615811] env[61868]: INFO nova.compute.manager [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] [instance: 4c98c21e-2918-4b55-b2b5-db14630a36ef] Took 1.03 seconds to deallocate network for instance. [ 652.759849] env[61868]: INFO nova.compute.manager [-] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Took 1.03 seconds to deallocate network for instance. [ 652.762805] env[61868]: DEBUG nova.compute.claims [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 652.762805] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 653.046347] env[61868]: DEBUG oslo_concurrency.lockutils [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.592s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 653.046872] env[61868]: DEBUG nova.compute.manager [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 653.049429] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.338s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 653.051010] env[61868]: INFO nova.compute.claims [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 653.557214] env[61868]: DEBUG nova.compute.utils [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 653.559841] env[61868]: DEBUG nova.compute.manager [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 653.561167] env[61868]: DEBUG nova.network.neutron [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 653.613194] env[61868]: DEBUG nova.policy [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd2c48e56b59b4b78b3676b1261d63360', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ccc6a8fff76843038eaef30e154e511f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 653.644066] env[61868]: INFO nova.scheduler.client.report [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] Deleted allocations for instance 4c98c21e-2918-4b55-b2b5-db14630a36ef [ 653.789049] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Acquiring lock "6391addb-b1a5-4e15-9bd2-531eb9956b42" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 653.789302] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Lock "6391addb-b1a5-4e15-9bd2-531eb9956b42" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 653.889438] env[61868]: DEBUG nova.network.neutron [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Successfully created port: 54acfe88-357f-46cb-b86d-5fdf61ac8f40 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 654.063654] env[61868]: DEBUG nova.compute.manager [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 654.151599] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e4afcc3c-c0f5-4c11-a381-1eb866579fcd tempest-ServersTestJSON-300143779 tempest-ServersTestJSON-300143779-project-member] Lock "4c98c21e-2918-4b55-b2b5-db14630a36ef" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 89.226s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 654.461302] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17702346-d3b9-4832-a398-6ac97129d6ca {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.469763] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc7aa3eb-844e-4656-b8b6-309f1cd1043e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.505289] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e02231f-b8d6-433b-a5a6-0bb1f572ab91 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.517057] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6290f1bd-c9c6-4e73-8f3b-5f9104accb89 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.534609] env[61868]: DEBUG nova.compute.provider_tree [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 654.656355] env[61868]: DEBUG nova.compute.manager [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 654.819127] env[61868]: DEBUG nova.compute.manager [req-baf7171e-153e-4e11-aad6-68dd6ef46ceb req-b27b82ad-c665-4d0a-ade9-2210b9d64919 service nova] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Received event network-changed-54acfe88-357f-46cb-b86d-5fdf61ac8f40 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 654.819127] env[61868]: DEBUG nova.compute.manager [req-baf7171e-153e-4e11-aad6-68dd6ef46ceb req-b27b82ad-c665-4d0a-ade9-2210b9d64919 service nova] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Refreshing instance network info cache due to event network-changed-54acfe88-357f-46cb-b86d-5fdf61ac8f40. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 654.819127] env[61868]: DEBUG oslo_concurrency.lockutils [req-baf7171e-153e-4e11-aad6-68dd6ef46ceb req-b27b82ad-c665-4d0a-ade9-2210b9d64919 service nova] Acquiring lock "refresh_cache-a8105842-1649-495d-b0c4-18ceda62bd57" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 654.819127] env[61868]: DEBUG oslo_concurrency.lockutils [req-baf7171e-153e-4e11-aad6-68dd6ef46ceb req-b27b82ad-c665-4d0a-ade9-2210b9d64919 service nova] Acquired lock "refresh_cache-a8105842-1649-495d-b0c4-18ceda62bd57" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.819127] env[61868]: DEBUG nova.network.neutron [req-baf7171e-153e-4e11-aad6-68dd6ef46ceb req-b27b82ad-c665-4d0a-ade9-2210b9d64919 service nova] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Refreshing network info cache for port 54acfe88-357f-46cb-b86d-5fdf61ac8f40 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 654.913817] env[61868]: ERROR nova.compute.manager [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 54acfe88-357f-46cb-b86d-5fdf61ac8f40, please check neutron logs for more information. [ 654.913817] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 654.913817] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 654.913817] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 654.913817] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 654.913817] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 654.913817] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 654.913817] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 654.913817] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.913817] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 654.913817] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.913817] env[61868]: ERROR nova.compute.manager raise self.value [ 654.913817] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 654.913817] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 654.913817] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.913817] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 654.914325] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.914325] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 654.914325] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 54acfe88-357f-46cb-b86d-5fdf61ac8f40, please check neutron logs for more information. [ 654.914325] env[61868]: ERROR nova.compute.manager [ 654.914325] env[61868]: Traceback (most recent call last): [ 654.914325] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 654.914325] env[61868]: listener.cb(fileno) [ 654.914325] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 654.914325] env[61868]: result = function(*args, **kwargs) [ 654.914325] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 654.914325] env[61868]: return func(*args, **kwargs) [ 654.914325] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 654.914325] env[61868]: raise e [ 654.914325] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 654.914325] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 654.914325] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 654.914325] env[61868]: created_port_ids = self._update_ports_for_instance( [ 654.914325] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 654.914325] env[61868]: with excutils.save_and_reraise_exception(): [ 654.914325] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.914325] env[61868]: self.force_reraise() [ 654.914325] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.914325] env[61868]: raise self.value [ 654.914325] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 654.914325] env[61868]: updated_port = self._update_port( [ 654.914325] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.914325] env[61868]: _ensure_no_port_binding_failure(port) [ 654.914325] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.914325] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 654.915141] env[61868]: nova.exception.PortBindingFailed: Binding failed for port 54acfe88-357f-46cb-b86d-5fdf61ac8f40, please check neutron logs for more information. [ 654.915141] env[61868]: Removing descriptor: 15 [ 655.039362] env[61868]: DEBUG nova.scheduler.client.report [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 655.079148] env[61868]: DEBUG nova.compute.manager [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 655.104497] env[61868]: DEBUG nova.virt.hardware [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 655.104497] env[61868]: DEBUG nova.virt.hardware [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 655.104497] env[61868]: DEBUG nova.virt.hardware [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 655.104784] env[61868]: DEBUG nova.virt.hardware [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 655.104784] env[61868]: DEBUG nova.virt.hardware [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 655.104784] env[61868]: DEBUG nova.virt.hardware [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 655.104784] env[61868]: DEBUG nova.virt.hardware [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 655.104784] env[61868]: DEBUG nova.virt.hardware [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 655.104914] env[61868]: DEBUG nova.virt.hardware [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 655.105056] env[61868]: DEBUG nova.virt.hardware [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 655.105302] env[61868]: DEBUG nova.virt.hardware [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 655.106151] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7056be1f-bea2-4a0e-bd7a-435a2ee013f1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.114371] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dbb954e-cf1e-453d-8d29-ccdc4ca94e4d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.128392] env[61868]: ERROR nova.compute.manager [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 54acfe88-357f-46cb-b86d-5fdf61ac8f40, please check neutron logs for more information. [ 655.128392] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Traceback (most recent call last): [ 655.128392] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 655.128392] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] yield resources [ 655.128392] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 655.128392] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] self.driver.spawn(context, instance, image_meta, [ 655.128392] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 655.128392] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] self._vmops.spawn(context, instance, image_meta, injected_files, [ 655.128392] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 655.128392] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] vm_ref = self.build_virtual_machine(instance, [ 655.128392] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 655.128788] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] vif_infos = vmwarevif.get_vif_info(self._session, [ 655.128788] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 655.128788] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] for vif in network_info: [ 655.128788] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 655.128788] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] return self._sync_wrapper(fn, *args, **kwargs) [ 655.128788] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 655.128788] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] self.wait() [ 655.128788] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 655.128788] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] self[:] = self._gt.wait() [ 655.128788] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 655.128788] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] return self._exit_event.wait() [ 655.128788] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 655.128788] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] current.throw(*self._exc) [ 655.129230] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 655.129230] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] result = function(*args, **kwargs) [ 655.129230] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 655.129230] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] return func(*args, **kwargs) [ 655.129230] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 655.129230] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] raise e [ 655.129230] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 655.129230] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] nwinfo = self.network_api.allocate_for_instance( [ 655.129230] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 655.129230] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] created_port_ids = self._update_ports_for_instance( [ 655.129230] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 655.129230] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] with excutils.save_and_reraise_exception(): [ 655.129230] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 655.129596] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] self.force_reraise() [ 655.129596] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 655.129596] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] raise self.value [ 655.129596] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 655.129596] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] updated_port = self._update_port( [ 655.129596] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 655.129596] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] _ensure_no_port_binding_failure(port) [ 655.129596] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 655.129596] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] raise exception.PortBindingFailed(port_id=port['id']) [ 655.129596] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] nova.exception.PortBindingFailed: Binding failed for port 54acfe88-357f-46cb-b86d-5fdf61ac8f40, please check neutron logs for more information. [ 655.129596] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] [ 655.129596] env[61868]: INFO nova.compute.manager [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Terminating instance [ 655.130637] env[61868]: DEBUG oslo_concurrency.lockutils [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Acquiring lock "refresh_cache-a8105842-1649-495d-b0c4-18ceda62bd57" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 655.175406] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 655.336598] env[61868]: DEBUG nova.network.neutron [req-baf7171e-153e-4e11-aad6-68dd6ef46ceb req-b27b82ad-c665-4d0a-ade9-2210b9d64919 service nova] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.389645] env[61868]: DEBUG nova.network.neutron [req-baf7171e-153e-4e11-aad6-68dd6ef46ceb req-b27b82ad-c665-4d0a-ade9-2210b9d64919 service nova] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.547018] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.495s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 655.547018] env[61868]: DEBUG nova.compute.manager [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 655.547454] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.705s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 655.554203] env[61868]: INFO nova.compute.claims [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 655.892506] env[61868]: DEBUG oslo_concurrency.lockutils [req-baf7171e-153e-4e11-aad6-68dd6ef46ceb req-b27b82ad-c665-4d0a-ade9-2210b9d64919 service nova] Releasing lock "refresh_cache-a8105842-1649-495d-b0c4-18ceda62bd57" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 655.894506] env[61868]: DEBUG oslo_concurrency.lockutils [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Acquired lock "refresh_cache-a8105842-1649-495d-b0c4-18ceda62bd57" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.894506] env[61868]: DEBUG nova.network.neutron [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 656.058215] env[61868]: DEBUG nova.compute.utils [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 656.062478] env[61868]: DEBUG nova.compute.manager [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 656.062478] env[61868]: DEBUG nova.network.neutron [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 656.132020] env[61868]: DEBUG nova.policy [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b620416fde09466097269120fc314bb6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6c9fd31be0044a4481c5efd882ccede1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 656.418074] env[61868]: DEBUG nova.network.neutron [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 656.441591] env[61868]: DEBUG nova.network.neutron [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Successfully created port: 59d1a6f1-3a24-4ce3-b41d-442fff832e84 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 656.504205] env[61868]: DEBUG nova.network.neutron [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.563064] env[61868]: DEBUG nova.compute.manager [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 656.961409] env[61868]: DEBUG nova.compute.manager [req-95736e6b-5801-43e8-b465-86ed79d42107 req-cf5853be-d2ca-4306-816a-656b774a8565 service nova] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Received event network-vif-deleted-54acfe88-357f-46cb-b86d-5fdf61ac8f40 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 657.009215] env[61868]: DEBUG oslo_concurrency.lockutils [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Releasing lock "refresh_cache-a8105842-1649-495d-b0c4-18ceda62bd57" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.010121] env[61868]: DEBUG nova.compute.manager [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 657.010121] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 657.010121] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dd71c9d5-0ce0-448d-a715-ecec0e6202b7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.025532] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e24ef698-51ab-432b-a63b-dd79b05b8edf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.039559] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b46b9cf-14af-47ab-8279-b2bd221ae578 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.047385] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-059c5d73-5089-40c2-8559-51ef7f17c6a4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.054407] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a8105842-1649-495d-b0c4-18ceda62bd57 could not be found. [ 657.054463] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 657.058019] env[61868]: INFO nova.compute.manager [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Took 0.04 seconds to destroy the instance on the hypervisor. [ 657.058019] env[61868]: DEBUG oslo.service.loopingcall [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 657.058019] env[61868]: DEBUG nova.compute.manager [-] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 657.058019] env[61868]: DEBUG nova.network.neutron [-] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 657.085514] env[61868]: DEBUG nova.network.neutron [-] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 657.086510] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3461b213-121b-467b-a7f2-00bf11e4df9c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.095248] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-774b15f1-01d6-4ac5-bbf5-41b1d216a545 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.109406] env[61868]: DEBUG nova.compute.provider_tree [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 657.591259] env[61868]: DEBUG nova.compute.manager [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 657.593369] env[61868]: DEBUG nova.network.neutron [-] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.613411] env[61868]: DEBUG nova.scheduler.client.report [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 657.626587] env[61868]: DEBUG nova.virt.hardware [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 657.626836] env[61868]: DEBUG nova.virt.hardware [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 657.626973] env[61868]: DEBUG nova.virt.hardware [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 657.627181] env[61868]: DEBUG nova.virt.hardware [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 657.627412] env[61868]: DEBUG nova.virt.hardware [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 657.627579] env[61868]: DEBUG nova.virt.hardware [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 657.627783] env[61868]: DEBUG nova.virt.hardware [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 657.627937] env[61868]: DEBUG nova.virt.hardware [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 657.628286] env[61868]: DEBUG nova.virt.hardware [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 657.628475] env[61868]: DEBUG nova.virt.hardware [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 657.628651] env[61868]: DEBUG nova.virt.hardware [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 657.630205] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-690ba6e7-73bc-445a-8dd7-d7915a4781e3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.633567] env[61868]: ERROR nova.compute.manager [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 59d1a6f1-3a24-4ce3-b41d-442fff832e84, please check neutron logs for more information. [ 657.633567] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 657.633567] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 657.633567] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 657.633567] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 657.633567] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 657.633567] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 657.633567] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 657.633567] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.633567] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 657.633567] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.633567] env[61868]: ERROR nova.compute.manager raise self.value [ 657.633567] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 657.633567] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 657.633567] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.633567] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 657.634029] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.634029] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 657.634029] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 59d1a6f1-3a24-4ce3-b41d-442fff832e84, please check neutron logs for more information. [ 657.634029] env[61868]: ERROR nova.compute.manager [ 657.634029] env[61868]: Traceback (most recent call last): [ 657.634029] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 657.634029] env[61868]: listener.cb(fileno) [ 657.634029] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 657.634029] env[61868]: result = function(*args, **kwargs) [ 657.634029] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 657.634029] env[61868]: return func(*args, **kwargs) [ 657.634029] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 657.634029] env[61868]: raise e [ 657.634029] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 657.634029] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 657.634029] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 657.634029] env[61868]: created_port_ids = self._update_ports_for_instance( [ 657.634029] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 657.634029] env[61868]: with excutils.save_and_reraise_exception(): [ 657.634029] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.634029] env[61868]: self.force_reraise() [ 657.634029] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.634029] env[61868]: raise self.value [ 657.634029] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 657.634029] env[61868]: updated_port = self._update_port( [ 657.634029] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.634029] env[61868]: _ensure_no_port_binding_failure(port) [ 657.634029] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.634029] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 657.634830] env[61868]: nova.exception.PortBindingFailed: Binding failed for port 59d1a6f1-3a24-4ce3-b41d-442fff832e84, please check neutron logs for more information. [ 657.634830] env[61868]: Removing descriptor: 15 [ 657.639652] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6bb5666-04a4-4fd1-aa22-5ca322fe6ae4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.655613] env[61868]: ERROR nova.compute.manager [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 59d1a6f1-3a24-4ce3-b41d-442fff832e84, please check neutron logs for more information. [ 657.655613] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Traceback (most recent call last): [ 657.655613] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 657.655613] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] yield resources [ 657.655613] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 657.655613] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] self.driver.spawn(context, instance, image_meta, [ 657.655613] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 657.655613] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] self._vmops.spawn(context, instance, image_meta, injected_files, [ 657.655613] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 657.655613] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] vm_ref = self.build_virtual_machine(instance, [ 657.655613] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 657.655997] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] vif_infos = vmwarevif.get_vif_info(self._session, [ 657.655997] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 657.655997] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] for vif in network_info: [ 657.655997] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 657.655997] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] return self._sync_wrapper(fn, *args, **kwargs) [ 657.655997] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 657.655997] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] self.wait() [ 657.655997] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 657.655997] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] self[:] = self._gt.wait() [ 657.655997] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 657.655997] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] return self._exit_event.wait() [ 657.655997] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 657.655997] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] current.throw(*self._exc) [ 657.656512] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 657.656512] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] result = function(*args, **kwargs) [ 657.656512] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 657.656512] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] return func(*args, **kwargs) [ 657.656512] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 657.656512] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] raise e [ 657.656512] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 657.656512] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] nwinfo = self.network_api.allocate_for_instance( [ 657.656512] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 657.656512] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] created_port_ids = self._update_ports_for_instance( [ 657.656512] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 657.656512] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] with excutils.save_and_reraise_exception(): [ 657.656512] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.656876] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] self.force_reraise() [ 657.656876] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.656876] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] raise self.value [ 657.656876] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 657.656876] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] updated_port = self._update_port( [ 657.656876] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.656876] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] _ensure_no_port_binding_failure(port) [ 657.656876] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.656876] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] raise exception.PortBindingFailed(port_id=port['id']) [ 657.656876] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] nova.exception.PortBindingFailed: Binding failed for port 59d1a6f1-3a24-4ce3-b41d-442fff832e84, please check neutron logs for more information. [ 657.656876] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] [ 657.656876] env[61868]: INFO nova.compute.manager [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Terminating instance [ 657.658277] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] Acquiring lock "refresh_cache-2d1b59b1-504c-487f-93ac-642c490d6546" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 657.658277] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] Acquired lock "refresh_cache-2d1b59b1-504c-487f-93ac-642c490d6546" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.658451] env[61868]: DEBUG nova.network.neutron [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 658.095695] env[61868]: INFO nova.compute.manager [-] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Took 1.04 seconds to deallocate network for instance. [ 658.098451] env[61868]: DEBUG nova.compute.claims [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 658.098632] env[61868]: DEBUG oslo_concurrency.lockutils [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 658.117706] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.570s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 658.118466] env[61868]: DEBUG nova.compute.manager [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 658.120724] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.244s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 658.122610] env[61868]: INFO nova.compute.claims [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 658.176789] env[61868]: DEBUG nova.network.neutron [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 658.270760] env[61868]: DEBUG nova.network.neutron [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.626593] env[61868]: DEBUG nova.compute.utils [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 658.629455] env[61868]: DEBUG nova.compute.manager [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 658.629628] env[61868]: DEBUG nova.network.neutron [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 658.672691] env[61868]: DEBUG nova.policy [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8e81e032784e4af78e8852b504c0ed67', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aa990fe138684572a856e80b8027a2fa', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 658.775181] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] Releasing lock "refresh_cache-2d1b59b1-504c-487f-93ac-642c490d6546" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 658.775618] env[61868]: DEBUG nova.compute.manager [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 658.775811] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 658.776115] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6d79f1a4-71ae-44fc-9e07-30d9e79180c2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.785110] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51563833-9dab-4c9d-9e7a-ef3fe1485d4d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.809173] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2d1b59b1-504c-487f-93ac-642c490d6546 could not be found. [ 658.809173] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 658.809173] env[61868]: INFO nova.compute.manager [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Took 0.03 seconds to destroy the instance on the hypervisor. [ 658.809173] env[61868]: DEBUG oslo.service.loopingcall [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 658.809173] env[61868]: DEBUG nova.compute.manager [-] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 658.809173] env[61868]: DEBUG nova.network.neutron [-] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 658.839043] env[61868]: DEBUG nova.network.neutron [-] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 658.967817] env[61868]: DEBUG nova.network.neutron [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Successfully created port: c7376ab9-07e7-4fe6-beac-650b9087fd7e {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 659.007210] env[61868]: DEBUG nova.compute.manager [req-da06208f-9850-4a09-88e9-e921aba7d773 req-5f436205-1b91-406f-8b0a-400500d1987c service nova] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Received event network-changed-59d1a6f1-3a24-4ce3-b41d-442fff832e84 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 659.007413] env[61868]: DEBUG nova.compute.manager [req-da06208f-9850-4a09-88e9-e921aba7d773 req-5f436205-1b91-406f-8b0a-400500d1987c service nova] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Refreshing instance network info cache due to event network-changed-59d1a6f1-3a24-4ce3-b41d-442fff832e84. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 659.007624] env[61868]: DEBUG oslo_concurrency.lockutils [req-da06208f-9850-4a09-88e9-e921aba7d773 req-5f436205-1b91-406f-8b0a-400500d1987c service nova] Acquiring lock "refresh_cache-2d1b59b1-504c-487f-93ac-642c490d6546" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 659.007769] env[61868]: DEBUG oslo_concurrency.lockutils [req-da06208f-9850-4a09-88e9-e921aba7d773 req-5f436205-1b91-406f-8b0a-400500d1987c service nova] Acquired lock "refresh_cache-2d1b59b1-504c-487f-93ac-642c490d6546" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.007913] env[61868]: DEBUG nova.network.neutron [req-da06208f-9850-4a09-88e9-e921aba7d773 req-5f436205-1b91-406f-8b0a-400500d1987c service nova] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Refreshing network info cache for port 59d1a6f1-3a24-4ce3-b41d-442fff832e84 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 659.136825] env[61868]: DEBUG nova.compute.manager [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 659.340815] env[61868]: DEBUG nova.network.neutron [-] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.507962] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d761ccb-68cd-4e55-99ff-7ad1780e3965 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.516655] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f191380-e224-4ca5-9878-deec01f231a3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.545290] env[61868]: DEBUG nova.network.neutron [req-da06208f-9850-4a09-88e9-e921aba7d773 req-5f436205-1b91-406f-8b0a-400500d1987c service nova] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 659.547360] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fdeeda0-8897-4226-a6f8-a7d8471615be {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.554859] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7634263f-c633-4496-a609-1389b5e353c9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.567441] env[61868]: DEBUG nova.compute.provider_tree [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 659.669751] env[61868]: DEBUG nova.network.neutron [req-da06208f-9850-4a09-88e9-e921aba7d773 req-5f436205-1b91-406f-8b0a-400500d1987c service nova] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.846617] env[61868]: INFO nova.compute.manager [-] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Took 1.04 seconds to deallocate network for instance. [ 659.848522] env[61868]: DEBUG nova.compute.claims [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 659.848662] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 659.878255] env[61868]: ERROR nova.compute.manager [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c7376ab9-07e7-4fe6-beac-650b9087fd7e, please check neutron logs for more information. [ 659.878255] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 659.878255] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 659.878255] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 659.878255] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 659.878255] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 659.878255] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 659.878255] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 659.878255] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 659.878255] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 659.878255] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 659.878255] env[61868]: ERROR nova.compute.manager raise self.value [ 659.878255] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 659.878255] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 659.878255] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 659.878255] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 659.878689] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 659.878689] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 659.878689] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c7376ab9-07e7-4fe6-beac-650b9087fd7e, please check neutron logs for more information. [ 659.878689] env[61868]: ERROR nova.compute.manager [ 659.878689] env[61868]: Traceback (most recent call last): [ 659.878689] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 659.878689] env[61868]: listener.cb(fileno) [ 659.878689] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 659.878689] env[61868]: result = function(*args, **kwargs) [ 659.878689] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 659.878689] env[61868]: return func(*args, **kwargs) [ 659.878689] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 659.878689] env[61868]: raise e [ 659.878689] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 659.878689] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 659.878689] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 659.878689] env[61868]: created_port_ids = self._update_ports_for_instance( [ 659.878689] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 659.878689] env[61868]: with excutils.save_and_reraise_exception(): [ 659.878689] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 659.878689] env[61868]: self.force_reraise() [ 659.878689] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 659.878689] env[61868]: raise self.value [ 659.878689] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 659.878689] env[61868]: updated_port = self._update_port( [ 659.878689] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 659.878689] env[61868]: _ensure_no_port_binding_failure(port) [ 659.878689] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 659.878689] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 659.879479] env[61868]: nova.exception.PortBindingFailed: Binding failed for port c7376ab9-07e7-4fe6-beac-650b9087fd7e, please check neutron logs for more information. [ 659.879479] env[61868]: Removing descriptor: 15 [ 660.071056] env[61868]: DEBUG nova.scheduler.client.report [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 660.146507] env[61868]: DEBUG nova.compute.manager [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 660.172803] env[61868]: DEBUG nova.virt.hardware [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 660.173083] env[61868]: DEBUG nova.virt.hardware [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 660.173308] env[61868]: DEBUG nova.virt.hardware [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 660.173461] env[61868]: DEBUG nova.virt.hardware [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 660.173607] env[61868]: DEBUG nova.virt.hardware [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 660.173751] env[61868]: DEBUG nova.virt.hardware [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 660.173956] env[61868]: DEBUG nova.virt.hardware [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 660.174128] env[61868]: DEBUG nova.virt.hardware [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 660.174297] env[61868]: DEBUG nova.virt.hardware [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 660.174455] env[61868]: DEBUG nova.virt.hardware [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 660.174655] env[61868]: DEBUG nova.virt.hardware [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 660.175126] env[61868]: DEBUG oslo_concurrency.lockutils [req-da06208f-9850-4a09-88e9-e921aba7d773 req-5f436205-1b91-406f-8b0a-400500d1987c service nova] Releasing lock "refresh_cache-2d1b59b1-504c-487f-93ac-642c490d6546" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 660.175374] env[61868]: DEBUG nova.compute.manager [req-da06208f-9850-4a09-88e9-e921aba7d773 req-5f436205-1b91-406f-8b0a-400500d1987c service nova] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Received event network-vif-deleted-59d1a6f1-3a24-4ce3-b41d-442fff832e84 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 660.176289] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-656ca6ad-9c65-4877-814f-d6dbb1475fa4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.184483] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9984fa9e-8cba-471d-b7fe-56e56dcfc02d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.199914] env[61868]: ERROR nova.compute.manager [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c7376ab9-07e7-4fe6-beac-650b9087fd7e, please check neutron logs for more information. [ 660.199914] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Traceback (most recent call last): [ 660.199914] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 660.199914] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] yield resources [ 660.199914] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 660.199914] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] self.driver.spawn(context, instance, image_meta, [ 660.199914] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 660.199914] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 660.199914] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 660.199914] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] vm_ref = self.build_virtual_machine(instance, [ 660.199914] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 660.200315] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] vif_infos = vmwarevif.get_vif_info(self._session, [ 660.200315] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 660.200315] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] for vif in network_info: [ 660.200315] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 660.200315] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] return self._sync_wrapper(fn, *args, **kwargs) [ 660.200315] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 660.200315] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] self.wait() [ 660.200315] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 660.200315] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] self[:] = self._gt.wait() [ 660.200315] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 660.200315] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] return self._exit_event.wait() [ 660.200315] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 660.200315] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] current.throw(*self._exc) [ 660.200657] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 660.200657] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] result = function(*args, **kwargs) [ 660.200657] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 660.200657] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] return func(*args, **kwargs) [ 660.200657] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 660.200657] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] raise e [ 660.200657] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.200657] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] nwinfo = self.network_api.allocate_for_instance( [ 660.200657] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 660.200657] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] created_port_ids = self._update_ports_for_instance( [ 660.200657] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 660.200657] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] with excutils.save_and_reraise_exception(): [ 660.200657] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.200976] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] self.force_reraise() [ 660.200976] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.200976] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] raise self.value [ 660.200976] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 660.200976] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] updated_port = self._update_port( [ 660.200976] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.200976] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] _ensure_no_port_binding_failure(port) [ 660.200976] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.200976] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] raise exception.PortBindingFailed(port_id=port['id']) [ 660.200976] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] nova.exception.PortBindingFailed: Binding failed for port c7376ab9-07e7-4fe6-beac-650b9087fd7e, please check neutron logs for more information. [ 660.200976] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] [ 660.200976] env[61868]: INFO nova.compute.manager [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Terminating instance [ 660.202641] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Acquiring lock "refresh_cache-3e970ecb-c842-4706-afd0-176d0088e1d3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.202836] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Acquired lock "refresh_cache-3e970ecb-c842-4706-afd0-176d0088e1d3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.203029] env[61868]: DEBUG nova.network.neutron [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 660.577049] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.455s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.577049] env[61868]: DEBUG nova.compute.manager [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 660.580069] env[61868]: DEBUG oslo_concurrency.lockutils [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 20.554s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.580299] env[61868]: DEBUG nova.objects.instance [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61868) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 660.719829] env[61868]: DEBUG nova.network.neutron [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 660.800879] env[61868]: DEBUG nova.network.neutron [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.030186] env[61868]: DEBUG nova.compute.manager [req-821b8a3b-1569-494f-a1d2-eb52e659477f req-946eb2bc-b5bc-4869-b63a-68113db6344c service nova] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Received event network-changed-c7376ab9-07e7-4fe6-beac-650b9087fd7e {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 661.030782] env[61868]: DEBUG nova.compute.manager [req-821b8a3b-1569-494f-a1d2-eb52e659477f req-946eb2bc-b5bc-4869-b63a-68113db6344c service nova] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Refreshing instance network info cache due to event network-changed-c7376ab9-07e7-4fe6-beac-650b9087fd7e. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 661.031008] env[61868]: DEBUG oslo_concurrency.lockutils [req-821b8a3b-1569-494f-a1d2-eb52e659477f req-946eb2bc-b5bc-4869-b63a-68113db6344c service nova] Acquiring lock "refresh_cache-3e970ecb-c842-4706-afd0-176d0088e1d3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 661.084403] env[61868]: DEBUG nova.compute.utils [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 661.088740] env[61868]: DEBUG nova.compute.manager [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 661.088740] env[61868]: DEBUG nova.network.neutron [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 661.127485] env[61868]: DEBUG nova.policy [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2b8b2470ec6441f38e7edb0b9ce69020', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f09bf553bf8b47e6846c59523d986edc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 661.304822] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Releasing lock "refresh_cache-3e970ecb-c842-4706-afd0-176d0088e1d3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.304822] env[61868]: DEBUG nova.compute.manager [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 661.304822] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 661.305150] env[61868]: DEBUG oslo_concurrency.lockutils [req-821b8a3b-1569-494f-a1d2-eb52e659477f req-946eb2bc-b5bc-4869-b63a-68113db6344c service nova] Acquired lock "refresh_cache-3e970ecb-c842-4706-afd0-176d0088e1d3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.305272] env[61868]: DEBUG nova.network.neutron [req-821b8a3b-1569-494f-a1d2-eb52e659477f req-946eb2bc-b5bc-4869-b63a-68113db6344c service nova] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Refreshing network info cache for port c7376ab9-07e7-4fe6-beac-650b9087fd7e {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 661.306330] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ecc17f1c-4ea1-49d7-af08-9acfae256875 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.321032] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eacdf01b-7097-44b1-ae5e-f20d36326667 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.344123] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3e970ecb-c842-4706-afd0-176d0088e1d3 could not be found. [ 661.344485] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 661.344764] env[61868]: INFO nova.compute.manager [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Took 0.04 seconds to destroy the instance on the hypervisor. [ 661.345114] env[61868]: DEBUG oslo.service.loopingcall [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 661.345440] env[61868]: DEBUG nova.compute.manager [-] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 661.345620] env[61868]: DEBUG nova.network.neutron [-] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 661.366557] env[61868]: DEBUG nova.network.neutron [-] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 661.408831] env[61868]: DEBUG nova.network.neutron [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Successfully created port: 7c54b4ab-e267-4ab2-856a-5b7ac33984c7 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 661.589143] env[61868]: DEBUG nova.compute.manager [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 661.592710] env[61868]: DEBUG oslo_concurrency.lockutils [None req-294db596-298b-4383-85e7-1dd8068e4e46 tempest-ServersAdmin275Test-91853274 tempest-ServersAdmin275Test-91853274-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.593816] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.603s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 661.830029] env[61868]: DEBUG nova.network.neutron [req-821b8a3b-1569-494f-a1d2-eb52e659477f req-946eb2bc-b5bc-4869-b63a-68113db6344c service nova] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 661.869299] env[61868]: DEBUG nova.network.neutron [-] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.944452] env[61868]: DEBUG nova.network.neutron [req-821b8a3b-1569-494f-a1d2-eb52e659477f req-946eb2bc-b5bc-4869-b63a-68113db6344c service nova] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.291137] env[61868]: ERROR nova.compute.manager [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7c54b4ab-e267-4ab2-856a-5b7ac33984c7, please check neutron logs for more information. [ 662.291137] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 662.291137] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 662.291137] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 662.291137] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 662.291137] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 662.291137] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 662.291137] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 662.291137] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 662.291137] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 662.291137] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 662.291137] env[61868]: ERROR nova.compute.manager raise self.value [ 662.291137] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 662.291137] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 662.291137] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 662.291137] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 662.291674] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 662.291674] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 662.291674] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7c54b4ab-e267-4ab2-856a-5b7ac33984c7, please check neutron logs for more information. [ 662.291674] env[61868]: ERROR nova.compute.manager [ 662.291674] env[61868]: Traceback (most recent call last): [ 662.291674] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 662.291674] env[61868]: listener.cb(fileno) [ 662.291674] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 662.291674] env[61868]: result = function(*args, **kwargs) [ 662.291674] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 662.291674] env[61868]: return func(*args, **kwargs) [ 662.291674] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 662.291674] env[61868]: raise e [ 662.291674] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 662.291674] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 662.291674] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 662.291674] env[61868]: created_port_ids = self._update_ports_for_instance( [ 662.291674] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 662.291674] env[61868]: with excutils.save_and_reraise_exception(): [ 662.291674] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 662.291674] env[61868]: self.force_reraise() [ 662.291674] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 662.291674] env[61868]: raise self.value [ 662.291674] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 662.291674] env[61868]: updated_port = self._update_port( [ 662.291674] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 662.291674] env[61868]: _ensure_no_port_binding_failure(port) [ 662.291674] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 662.291674] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 662.292461] env[61868]: nova.exception.PortBindingFailed: Binding failed for port 7c54b4ab-e267-4ab2-856a-5b7ac33984c7, please check neutron logs for more information. [ 662.292461] env[61868]: Removing descriptor: 15 [ 662.372679] env[61868]: INFO nova.compute.manager [-] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Took 1.03 seconds to deallocate network for instance. [ 662.375667] env[61868]: DEBUG nova.compute.claims [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 662.375838] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.448571] env[61868]: DEBUG oslo_concurrency.lockutils [req-821b8a3b-1569-494f-a1d2-eb52e659477f req-946eb2bc-b5bc-4869-b63a-68113db6344c service nova] Releasing lock "refresh_cache-3e970ecb-c842-4706-afd0-176d0088e1d3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 662.448874] env[61868]: DEBUG nova.compute.manager [req-821b8a3b-1569-494f-a1d2-eb52e659477f req-946eb2bc-b5bc-4869-b63a-68113db6344c service nova] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Received event network-vif-deleted-c7376ab9-07e7-4fe6-beac-650b9087fd7e {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 662.472712] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bde5da07-5b13-462e-86fe-0e3a691df9ef {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.480229] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7820ec08-3376-423b-bece-313182c3dee9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.508921] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-915c2fcd-624c-4fdb-bd61-8418542ef779 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.516288] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-890e6b94-15da-4412-a19e-61791d97b3d1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.529016] env[61868]: DEBUG nova.compute.provider_tree [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 662.599159] env[61868]: DEBUG nova.compute.manager [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 662.622824] env[61868]: DEBUG nova.virt.hardware [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 662.623069] env[61868]: DEBUG nova.virt.hardware [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 662.623228] env[61868]: DEBUG nova.virt.hardware [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 662.623406] env[61868]: DEBUG nova.virt.hardware [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 662.623548] env[61868]: DEBUG nova.virt.hardware [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 662.623688] env[61868]: DEBUG nova.virt.hardware [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 662.623887] env[61868]: DEBUG nova.virt.hardware [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 662.624051] env[61868]: DEBUG nova.virt.hardware [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 662.624215] env[61868]: DEBUG nova.virt.hardware [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 662.624374] env[61868]: DEBUG nova.virt.hardware [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 662.624540] env[61868]: DEBUG nova.virt.hardware [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 662.625362] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a04cccc-03de-4b35-945a-a1d616c684f5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.632631] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce2af61b-3137-4a87-8b1c-8f693f19a374 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.646702] env[61868]: ERROR nova.compute.manager [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7c54b4ab-e267-4ab2-856a-5b7ac33984c7, please check neutron logs for more information. [ 662.646702] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Traceback (most recent call last): [ 662.646702] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 662.646702] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] yield resources [ 662.646702] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 662.646702] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] self.driver.spawn(context, instance, image_meta, [ 662.646702] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 662.646702] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 662.646702] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 662.646702] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] vm_ref = self.build_virtual_machine(instance, [ 662.646702] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 662.647073] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] vif_infos = vmwarevif.get_vif_info(self._session, [ 662.647073] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 662.647073] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] for vif in network_info: [ 662.647073] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 662.647073] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] return self._sync_wrapper(fn, *args, **kwargs) [ 662.647073] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 662.647073] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] self.wait() [ 662.647073] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 662.647073] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] self[:] = self._gt.wait() [ 662.647073] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 662.647073] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] return self._exit_event.wait() [ 662.647073] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 662.647073] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] current.throw(*self._exc) [ 662.647434] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 662.647434] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] result = function(*args, **kwargs) [ 662.647434] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 662.647434] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] return func(*args, **kwargs) [ 662.647434] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 662.647434] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] raise e [ 662.647434] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 662.647434] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] nwinfo = self.network_api.allocate_for_instance( [ 662.647434] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 662.647434] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] created_port_ids = self._update_ports_for_instance( [ 662.647434] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 662.647434] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] with excutils.save_and_reraise_exception(): [ 662.647434] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 662.647799] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] self.force_reraise() [ 662.647799] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 662.647799] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] raise self.value [ 662.647799] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 662.647799] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] updated_port = self._update_port( [ 662.647799] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 662.647799] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] _ensure_no_port_binding_failure(port) [ 662.647799] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 662.647799] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] raise exception.PortBindingFailed(port_id=port['id']) [ 662.647799] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] nova.exception.PortBindingFailed: Binding failed for port 7c54b4ab-e267-4ab2-856a-5b7ac33984c7, please check neutron logs for more information. [ 662.647799] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] [ 662.647799] env[61868]: INFO nova.compute.manager [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Terminating instance [ 662.648913] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Acquiring lock "refresh_cache-85dbaf67-049c-4a90-a3dc-67e2decc1dbf" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 662.649086] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Acquired lock "refresh_cache-85dbaf67-049c-4a90-a3dc-67e2decc1dbf" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 662.649249] env[61868]: DEBUG nova.network.neutron [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 663.032110] env[61868]: DEBUG nova.scheduler.client.report [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 663.080957] env[61868]: DEBUG nova.compute.manager [req-6d2d7983-f6e8-4a75-aef4-4069f9fcb172 req-0702ddac-b8ae-4678-b7ef-533e5baa37f9 service nova] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Received event network-changed-7c54b4ab-e267-4ab2-856a-5b7ac33984c7 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 663.081247] env[61868]: DEBUG nova.compute.manager [req-6d2d7983-f6e8-4a75-aef4-4069f9fcb172 req-0702ddac-b8ae-4678-b7ef-533e5baa37f9 service nova] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Refreshing instance network info cache due to event network-changed-7c54b4ab-e267-4ab2-856a-5b7ac33984c7. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 663.081526] env[61868]: DEBUG oslo_concurrency.lockutils [req-6d2d7983-f6e8-4a75-aef4-4069f9fcb172 req-0702ddac-b8ae-4678-b7ef-533e5baa37f9 service nova] Acquiring lock "refresh_cache-85dbaf67-049c-4a90-a3dc-67e2decc1dbf" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 663.166113] env[61868]: DEBUG nova.network.neutron [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 663.247581] env[61868]: DEBUG nova.network.neutron [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.537093] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.943s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.537736] env[61868]: ERROR nova.compute.manager [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a4c26f3e-a6e6-45b9-b449-bf07b77ca99b, please check neutron logs for more information. [ 663.537736] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Traceback (most recent call last): [ 663.537736] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 663.537736] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] self.driver.spawn(context, instance, image_meta, [ 663.537736] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 663.537736] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] self._vmops.spawn(context, instance, image_meta, injected_files, [ 663.537736] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 663.537736] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] vm_ref = self.build_virtual_machine(instance, [ 663.537736] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 663.537736] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] vif_infos = vmwarevif.get_vif_info(self._session, [ 663.537736] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 663.538056] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] for vif in network_info: [ 663.538056] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 663.538056] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] return self._sync_wrapper(fn, *args, **kwargs) [ 663.538056] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 663.538056] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] self.wait() [ 663.538056] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 663.538056] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] self[:] = self._gt.wait() [ 663.538056] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 663.538056] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] return self._exit_event.wait() [ 663.538056] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 663.538056] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] result = hub.switch() [ 663.538056] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 663.538056] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] return self.greenlet.switch() [ 663.538369] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 663.538369] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] result = function(*args, **kwargs) [ 663.538369] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 663.538369] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] return func(*args, **kwargs) [ 663.538369] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 663.538369] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] raise e [ 663.538369] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 663.538369] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] nwinfo = self.network_api.allocate_for_instance( [ 663.538369] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 663.538369] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] created_port_ids = self._update_ports_for_instance( [ 663.538369] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 663.538369] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] with excutils.save_and_reraise_exception(): [ 663.538369] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.538714] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] self.force_reraise() [ 663.538714] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.538714] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] raise self.value [ 663.538714] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 663.538714] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] updated_port = self._update_port( [ 663.538714] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.538714] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] _ensure_no_port_binding_failure(port) [ 663.538714] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.538714] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] raise exception.PortBindingFailed(port_id=port['id']) [ 663.538714] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] nova.exception.PortBindingFailed: Binding failed for port a4c26f3e-a6e6-45b9-b449-bf07b77ca99b, please check neutron logs for more information. [ 663.538714] env[61868]: ERROR nova.compute.manager [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] [ 663.538973] env[61868]: DEBUG nova.compute.utils [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Binding failed for port a4c26f3e-a6e6-45b9-b449-bf07b77ca99b, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 663.539752] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.352s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 663.541221] env[61868]: INFO nova.compute.claims [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 663.543867] env[61868]: DEBUG nova.compute.manager [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Build of instance c46a8746-b4b2-46f8-b16a-a7ceb5ff4611 was re-scheduled: Binding failed for port a4c26f3e-a6e6-45b9-b449-bf07b77ca99b, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 663.544322] env[61868]: DEBUG nova.compute.manager [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 663.544536] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] Acquiring lock "refresh_cache-c46a8746-b4b2-46f8-b16a-a7ceb5ff4611" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 663.544682] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] Acquired lock "refresh_cache-c46a8746-b4b2-46f8-b16a-a7ceb5ff4611" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.544836] env[61868]: DEBUG nova.network.neutron [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 663.750746] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Releasing lock "refresh_cache-85dbaf67-049c-4a90-a3dc-67e2decc1dbf" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 663.751054] env[61868]: DEBUG nova.compute.manager [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 663.751257] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 663.751569] env[61868]: DEBUG oslo_concurrency.lockutils [req-6d2d7983-f6e8-4a75-aef4-4069f9fcb172 req-0702ddac-b8ae-4678-b7ef-533e5baa37f9 service nova] Acquired lock "refresh_cache-85dbaf67-049c-4a90-a3dc-67e2decc1dbf" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.751762] env[61868]: DEBUG nova.network.neutron [req-6d2d7983-f6e8-4a75-aef4-4069f9fcb172 req-0702ddac-b8ae-4678-b7ef-533e5baa37f9 service nova] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Refreshing network info cache for port 7c54b4ab-e267-4ab2-856a-5b7ac33984c7 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 663.753195] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-98f689cc-248a-43a5-aa47-01b16956ef2d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.762565] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66334bc4-9277-4d53-8683-92958a519a9d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.783919] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 85dbaf67-049c-4a90-a3dc-67e2decc1dbf could not be found. [ 663.784138] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 663.784548] env[61868]: INFO nova.compute.manager [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Took 0.03 seconds to destroy the instance on the hypervisor. [ 663.784548] env[61868]: DEBUG oslo.service.loopingcall [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 663.784741] env[61868]: DEBUG nova.compute.manager [-] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 663.784831] env[61868]: DEBUG nova.network.neutron [-] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 663.799345] env[61868]: DEBUG nova.network.neutron [-] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 664.066292] env[61868]: DEBUG nova.network.neutron [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 664.134495] env[61868]: DEBUG nova.network.neutron [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.271201] env[61868]: DEBUG nova.network.neutron [req-6d2d7983-f6e8-4a75-aef4-4069f9fcb172 req-0702ddac-b8ae-4678-b7ef-533e5baa37f9 service nova] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 664.302766] env[61868]: DEBUG nova.network.neutron [-] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.341622] env[61868]: DEBUG nova.network.neutron [req-6d2d7983-f6e8-4a75-aef4-4069f9fcb172 req-0702ddac-b8ae-4678-b7ef-533e5baa37f9 service nova] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.638354] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] Releasing lock "refresh_cache-c46a8746-b4b2-46f8-b16a-a7ceb5ff4611" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 664.638590] env[61868]: DEBUG nova.compute.manager [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 664.638779] env[61868]: DEBUG nova.compute.manager [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 664.638940] env[61868]: DEBUG nova.network.neutron [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 664.655254] env[61868]: DEBUG nova.network.neutron [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 664.804959] env[61868]: INFO nova.compute.manager [-] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Took 1.02 seconds to deallocate network for instance. [ 664.807247] env[61868]: DEBUG nova.compute.claims [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 664.807470] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.844440] env[61868]: DEBUG oslo_concurrency.lockutils [req-6d2d7983-f6e8-4a75-aef4-4069f9fcb172 req-0702ddac-b8ae-4678-b7ef-533e5baa37f9 service nova] Releasing lock "refresh_cache-85dbaf67-049c-4a90-a3dc-67e2decc1dbf" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 664.844684] env[61868]: DEBUG nova.compute.manager [req-6d2d7983-f6e8-4a75-aef4-4069f9fcb172 req-0702ddac-b8ae-4678-b7ef-533e5baa37f9 service nova] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Received event network-vif-deleted-7c54b4ab-e267-4ab2-856a-5b7ac33984c7 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 664.890254] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e961225e-3878-4cf4-b1a9-795277484710 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.897627] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db81bdd5-75f7-4386-b995-06bd2de73723 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.926529] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01fb8e8c-b0db-48fe-98ee-b64f5f1e7b82 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.933427] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-138cfbab-6203-4315-9060-d5d5e2fd9627 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.946440] env[61868]: DEBUG nova.compute.provider_tree [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 665.158269] env[61868]: DEBUG nova.network.neutron [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.449914] env[61868]: DEBUG nova.scheduler.client.report [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 665.661527] env[61868]: INFO nova.compute.manager [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] [instance: c46a8746-b4b2-46f8-b16a-a7ceb5ff4611] Took 1.02 seconds to deallocate network for instance. [ 665.956055] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.416s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 665.956055] env[61868]: DEBUG nova.compute.manager [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 665.959030] env[61868]: DEBUG oslo_concurrency.lockutils [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.591s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 665.959152] env[61868]: DEBUG nova.objects.instance [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Lazy-loading 'resources' on Instance uuid 7651f6d2-1825-46a1-9380-732b50abd3cc {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 666.460282] env[61868]: DEBUG nova.compute.utils [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 666.461685] env[61868]: DEBUG nova.compute.manager [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 666.461855] env[61868]: DEBUG nova.network.neutron [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: c979d105-12b7-485c-a136-55ffad02fe67] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 666.504063] env[61868]: DEBUG nova.policy [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8e81e032784e4af78e8852b504c0ed67', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aa990fe138684572a856e80b8027a2fa', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 666.691732] env[61868]: INFO nova.scheduler.client.report [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] Deleted allocations for instance c46a8746-b4b2-46f8-b16a-a7ceb5ff4611 [ 666.782224] env[61868]: DEBUG nova.network.neutron [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Successfully created port: ac679a13-bea7-4eeb-9517-68d1b5292e42 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 666.933913] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c352cffb-e04a-4226-88af-309b0834e2d6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.941561] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca32b388-73f0-47f6-962f-02c1392c6ec7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.978958] env[61868]: DEBUG nova.compute.manager [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 666.982378] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d1c0128-a553-4f7e-9dc3-ba1ba24070f0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.996270] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d71bc67-d959-450a-b190-f29be3ba5335 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.009650] env[61868]: DEBUG nova.compute.provider_tree [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 667.205988] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d4f8bf8-85cd-481a-8134-bc3ef00732d1 tempest-ImagesOneServerNegativeTestJSON-1780891644 tempest-ImagesOneServerNegativeTestJSON-1780891644-project-member] Lock "c46a8746-b4b2-46f8-b16a-a7ceb5ff4611" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 101.283s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.514814] env[61868]: DEBUG nova.scheduler.client.report [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 667.716752] env[61868]: DEBUG nova.compute.manager [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 667.917462] env[61868]: DEBUG nova.compute.manager [req-613d28dd-a0fa-4717-a6df-faeb0cde76c0 req-0e1934d9-414c-48be-9d3c-f47c3e0b1009 service nova] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Received event network-changed-ac679a13-bea7-4eeb-9517-68d1b5292e42 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 667.917462] env[61868]: DEBUG nova.compute.manager [req-613d28dd-a0fa-4717-a6df-faeb0cde76c0 req-0e1934d9-414c-48be-9d3c-f47c3e0b1009 service nova] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Refreshing instance network info cache due to event network-changed-ac679a13-bea7-4eeb-9517-68d1b5292e42. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 667.917462] env[61868]: DEBUG oslo_concurrency.lockutils [req-613d28dd-a0fa-4717-a6df-faeb0cde76c0 req-0e1934d9-414c-48be-9d3c-f47c3e0b1009 service nova] Acquiring lock "refresh_cache-c979d105-12b7-485c-a136-55ffad02fe67" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 667.917462] env[61868]: DEBUG oslo_concurrency.lockutils [req-613d28dd-a0fa-4717-a6df-faeb0cde76c0 req-0e1934d9-414c-48be-9d3c-f47c3e0b1009 service nova] Acquired lock "refresh_cache-c979d105-12b7-485c-a136-55ffad02fe67" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.917462] env[61868]: DEBUG nova.network.neutron [req-613d28dd-a0fa-4717-a6df-faeb0cde76c0 req-0e1934d9-414c-48be-9d3c-f47c3e0b1009 service nova] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Refreshing network info cache for port ac679a13-bea7-4eeb-9517-68d1b5292e42 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 667.997158] env[61868]: DEBUG nova.compute.manager [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 668.023192] env[61868]: DEBUG oslo_concurrency.lockutils [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.061s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 668.023192] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 19.953s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.023192] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 668.023192] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61868) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 668.024715] env[61868]: DEBUG oslo_concurrency.lockutils [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.493s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.031591] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fd9e02f-4835-490c-af7a-a5cfa71b8012 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.042328] env[61868]: DEBUG nova.virt.hardware [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 668.042760] env[61868]: DEBUG nova.virt.hardware [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 668.043036] env[61868]: DEBUG nova.virt.hardware [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 668.043322] env[61868]: DEBUG nova.virt.hardware [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 668.043574] env[61868]: DEBUG nova.virt.hardware [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 668.043953] env[61868]: DEBUG nova.virt.hardware [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 668.044381] env[61868]: DEBUG nova.virt.hardware [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 668.044637] env[61868]: DEBUG nova.virt.hardware [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 668.047049] env[61868]: DEBUG nova.virt.hardware [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 668.047049] env[61868]: DEBUG nova.virt.hardware [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 668.047049] env[61868]: DEBUG nova.virt.hardware [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 668.047049] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e7423e7-cf98-4244-8418-d30aeccb27f2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.050773] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24c1e35d-6480-446f-8f90-6f32a8ce6ea1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.071660] env[61868]: INFO nova.scheduler.client.report [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Deleted allocations for instance 7651f6d2-1825-46a1-9380-732b50abd3cc [ 668.076468] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1c07956-289d-4bd9-8fa2-f5b8043ad86f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.081521] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b704e64-36f7-4da6-95f6-2590b8680926 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.100904] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9e29a56-4f61-4502-bf8b-6f38a9f2bd28 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.135593] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181531MB free_disk=134GB free_vcpus=48 pci_devices=None {{(pid=61868) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 668.135755] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.236446] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.250046] env[61868]: ERROR nova.compute.manager [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ac679a13-bea7-4eeb-9517-68d1b5292e42, please check neutron logs for more information. [ 668.250046] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 668.250046] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 668.250046] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 668.250046] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 668.250046] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 668.250046] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 668.250046] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 668.250046] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 668.250046] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 668.250046] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 668.250046] env[61868]: ERROR nova.compute.manager raise self.value [ 668.250046] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 668.250046] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 668.250046] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 668.250046] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 668.250529] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 668.250529] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 668.250529] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ac679a13-bea7-4eeb-9517-68d1b5292e42, please check neutron logs for more information. [ 668.250529] env[61868]: ERROR nova.compute.manager [ 668.250529] env[61868]: Traceback (most recent call last): [ 668.250529] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 668.250529] env[61868]: listener.cb(fileno) [ 668.250529] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 668.250529] env[61868]: result = function(*args, **kwargs) [ 668.250529] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 668.250529] env[61868]: return func(*args, **kwargs) [ 668.250529] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 668.250529] env[61868]: raise e [ 668.250529] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 668.250529] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 668.250529] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 668.250529] env[61868]: created_port_ids = self._update_ports_for_instance( [ 668.250529] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 668.250529] env[61868]: with excutils.save_and_reraise_exception(): [ 668.250529] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 668.250529] env[61868]: self.force_reraise() [ 668.250529] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 668.250529] env[61868]: raise self.value [ 668.250529] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 668.250529] env[61868]: updated_port = self._update_port( [ 668.250529] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 668.250529] env[61868]: _ensure_no_port_binding_failure(port) [ 668.250529] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 668.250529] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 668.251302] env[61868]: nova.exception.PortBindingFailed: Binding failed for port ac679a13-bea7-4eeb-9517-68d1b5292e42, please check neutron logs for more information. [ 668.251302] env[61868]: Removing descriptor: 15 [ 668.251302] env[61868]: ERROR nova.compute.manager [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ac679a13-bea7-4eeb-9517-68d1b5292e42, please check neutron logs for more information. [ 668.251302] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] Traceback (most recent call last): [ 668.251302] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 668.251302] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] yield resources [ 668.251302] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 668.251302] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] self.driver.spawn(context, instance, image_meta, [ 668.251302] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 668.251302] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] self._vmops.spawn(context, instance, image_meta, injected_files, [ 668.251302] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 668.251302] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] vm_ref = self.build_virtual_machine(instance, [ 668.251628] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 668.251628] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] vif_infos = vmwarevif.get_vif_info(self._session, [ 668.251628] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 668.251628] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] for vif in network_info: [ 668.251628] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 668.251628] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] return self._sync_wrapper(fn, *args, **kwargs) [ 668.251628] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 668.251628] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] self.wait() [ 668.251628] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 668.251628] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] self[:] = self._gt.wait() [ 668.251628] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 668.251628] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] return self._exit_event.wait() [ 668.251628] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 668.251942] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] result = hub.switch() [ 668.251942] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 668.251942] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] return self.greenlet.switch() [ 668.251942] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 668.251942] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] result = function(*args, **kwargs) [ 668.251942] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 668.251942] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] return func(*args, **kwargs) [ 668.251942] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 668.251942] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] raise e [ 668.251942] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 668.251942] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] nwinfo = self.network_api.allocate_for_instance( [ 668.251942] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 668.251942] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] created_port_ids = self._update_ports_for_instance( [ 668.252268] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 668.252268] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] with excutils.save_and_reraise_exception(): [ 668.252268] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 668.252268] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] self.force_reraise() [ 668.252268] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 668.252268] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] raise self.value [ 668.252268] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 668.252268] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] updated_port = self._update_port( [ 668.252268] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 668.252268] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] _ensure_no_port_binding_failure(port) [ 668.252268] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 668.252268] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] raise exception.PortBindingFailed(port_id=port['id']) [ 668.252593] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] nova.exception.PortBindingFailed: Binding failed for port ac679a13-bea7-4eeb-9517-68d1b5292e42, please check neutron logs for more information. [ 668.252593] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] [ 668.252593] env[61868]: INFO nova.compute.manager [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Terminating instance [ 668.253621] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Acquiring lock "refresh_cache-c979d105-12b7-485c-a136-55ffad02fe67" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 668.441436] env[61868]: DEBUG nova.network.neutron [req-613d28dd-a0fa-4717-a6df-faeb0cde76c0 req-0e1934d9-414c-48be-9d3c-f47c3e0b1009 service nova] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 668.548940] env[61868]: DEBUG nova.network.neutron [req-613d28dd-a0fa-4717-a6df-faeb0cde76c0 req-0e1934d9-414c-48be-9d3c-f47c3e0b1009 service nova] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.585734] env[61868]: DEBUG oslo_concurrency.lockutils [None req-800f8122-01ab-461b-a76c-1dc109a7f9eb tempest-ServersAdmin275Test-192996573 tempest-ServersAdmin275Test-192996573-project-member] Lock "7651f6d2-1825-46a1-9380-732b50abd3cc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.227s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 668.945274] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80d8ae1e-e6d6-4a48-8712-2963815c8b35 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.953865] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-119fec03-b13e-44bf-88e1-43806d8ae05a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.986047] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-579b8ccf-235b-41c3-84bc-313592ac29d3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.995305] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efcd8313-7db1-4a8d-ba8f-e2d87ced6b7e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.011337] env[61868]: DEBUG nova.compute.provider_tree [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 669.051580] env[61868]: DEBUG oslo_concurrency.lockutils [req-613d28dd-a0fa-4717-a6df-faeb0cde76c0 req-0e1934d9-414c-48be-9d3c-f47c3e0b1009 service nova] Releasing lock "refresh_cache-c979d105-12b7-485c-a136-55ffad02fe67" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 669.052717] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Acquired lock "refresh_cache-c979d105-12b7-485c-a136-55ffad02fe67" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.052717] env[61868]: DEBUG nova.network.neutron [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 669.515212] env[61868]: DEBUG nova.scheduler.client.report [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 669.572324] env[61868]: DEBUG nova.network.neutron [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 669.656720] env[61868]: DEBUG nova.network.neutron [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.959826] env[61868]: DEBUG nova.compute.manager [req-58e0cb5e-65a7-422d-9ec8-59c60a4a4faf req-ca834b42-85e4-4f6d-a3f5-765888fa7ba4 service nova] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Received event network-vif-deleted-ac679a13-bea7-4eeb-9517-68d1b5292e42 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 670.020149] env[61868]: DEBUG oslo_concurrency.lockutils [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.995s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.020875] env[61868]: ERROR nova.compute.manager [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6cdafb9f-87e5-4c0a-8bd1-1fcec1a98833, please check neutron logs for more information. [ 670.020875] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Traceback (most recent call last): [ 670.020875] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 670.020875] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] self.driver.spawn(context, instance, image_meta, [ 670.020875] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 670.020875] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 670.020875] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 670.020875] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] vm_ref = self.build_virtual_machine(instance, [ 670.020875] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 670.020875] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] vif_infos = vmwarevif.get_vif_info(self._session, [ 670.020875] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 670.021183] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] for vif in network_info: [ 670.021183] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 670.021183] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] return self._sync_wrapper(fn, *args, **kwargs) [ 670.021183] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 670.021183] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] self.wait() [ 670.021183] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 670.021183] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] self[:] = self._gt.wait() [ 670.021183] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 670.021183] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] return self._exit_event.wait() [ 670.021183] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 670.021183] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] result = hub.switch() [ 670.021183] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 670.021183] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] return self.greenlet.switch() [ 670.021687] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 670.021687] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] result = function(*args, **kwargs) [ 670.021687] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 670.021687] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] return func(*args, **kwargs) [ 670.021687] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 670.021687] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] raise e [ 670.021687] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 670.021687] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] nwinfo = self.network_api.allocate_for_instance( [ 670.021687] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 670.021687] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] created_port_ids = self._update_ports_for_instance( [ 670.021687] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 670.021687] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] with excutils.save_and_reraise_exception(): [ 670.021687] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.021997] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] self.force_reraise() [ 670.021997] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.021997] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] raise self.value [ 670.021997] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 670.021997] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] updated_port = self._update_port( [ 670.021997] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.021997] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] _ensure_no_port_binding_failure(port) [ 670.021997] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.021997] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] raise exception.PortBindingFailed(port_id=port['id']) [ 670.021997] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] nova.exception.PortBindingFailed: Binding failed for port 6cdafb9f-87e5-4c0a-8bd1-1fcec1a98833, please check neutron logs for more information. [ 670.021997] env[61868]: ERROR nova.compute.manager [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] [ 670.022276] env[61868]: DEBUG nova.compute.utils [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Binding failed for port 6cdafb9f-87e5-4c0a-8bd1-1fcec1a98833, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 670.022837] env[61868]: DEBUG oslo_concurrency.lockutils [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.518s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.024439] env[61868]: INFO nova.compute.claims [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 670.027109] env[61868]: DEBUG nova.compute.manager [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Build of instance 1586080c-3406-47f0-bbb9-0c50082eaf8d was re-scheduled: Binding failed for port 6cdafb9f-87e5-4c0a-8bd1-1fcec1a98833, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 670.027439] env[61868]: DEBUG nova.compute.manager [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 670.027746] env[61868]: DEBUG oslo_concurrency.lockutils [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] Acquiring lock "refresh_cache-1586080c-3406-47f0-bbb9-0c50082eaf8d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 670.027824] env[61868]: DEBUG oslo_concurrency.lockutils [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] Acquired lock "refresh_cache-1586080c-3406-47f0-bbb9-0c50082eaf8d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.027941] env[61868]: DEBUG nova.network.neutron [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 670.160988] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Releasing lock "refresh_cache-c979d105-12b7-485c-a136-55ffad02fe67" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 670.161519] env[61868]: DEBUG nova.compute.manager [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 670.161780] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 670.162095] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-353626af-0d52-48a2-9ddd-706de298d378 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.174733] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61df2111-deee-4e2c-9125-9c99244b1a80 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.204235] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c979d105-12b7-485c-a136-55ffad02fe67 could not be found. [ 670.204522] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 670.204736] env[61868]: INFO nova.compute.manager [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Took 0.04 seconds to destroy the instance on the hypervisor. [ 670.204961] env[61868]: DEBUG oslo.service.loopingcall [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 670.205220] env[61868]: DEBUG nova.compute.manager [-] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 670.205316] env[61868]: DEBUG nova.network.neutron [-] [instance: c979d105-12b7-485c-a136-55ffad02fe67] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 670.227880] env[61868]: DEBUG nova.network.neutron [-] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 670.550558] env[61868]: DEBUG nova.network.neutron [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 670.627557] env[61868]: DEBUG nova.network.neutron [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.730969] env[61868]: DEBUG nova.network.neutron [-] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.132809] env[61868]: DEBUG oslo_concurrency.lockutils [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] Releasing lock "refresh_cache-1586080c-3406-47f0-bbb9-0c50082eaf8d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 671.133164] env[61868]: DEBUG nova.compute.manager [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 671.133423] env[61868]: DEBUG nova.compute.manager [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 671.133641] env[61868]: DEBUG nova.network.neutron [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 671.161213] env[61868]: DEBUG nova.network.neutron [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 671.239239] env[61868]: INFO nova.compute.manager [-] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Took 1.03 seconds to deallocate network for instance. [ 671.242989] env[61868]: DEBUG nova.compute.claims [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 671.242989] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 671.591182] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c8fdb91-95f8-4364-8c63-0ec5ce5b44c4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.598985] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-694fc3e6-f0f2-4902-954a-93ec7ddfd576 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.636728] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d61f11b-b940-44dc-8e40-d7d26640325b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.648041] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04cce6af-66f0-4a0e-a7f7-5daea714a284 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.665859] env[61868]: DEBUG nova.compute.provider_tree [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 671.667379] env[61868]: DEBUG nova.network.neutron [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.169718] env[61868]: DEBUG nova.scheduler.client.report [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 672.175397] env[61868]: INFO nova.compute.manager [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] [instance: 1586080c-3406-47f0-bbb9-0c50082eaf8d] Took 1.04 seconds to deallocate network for instance. [ 672.681883] env[61868]: DEBUG oslo_concurrency.lockutils [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.659s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 672.682919] env[61868]: DEBUG nova.compute.manager [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 672.692020] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.928s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.187953] env[61868]: DEBUG nova.compute.utils [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 673.189442] env[61868]: DEBUG nova.compute.manager [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 673.191723] env[61868]: DEBUG nova.network.neutron [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 673.225044] env[61868]: INFO nova.scheduler.client.report [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] Deleted allocations for instance 1586080c-3406-47f0-bbb9-0c50082eaf8d [ 673.261831] env[61868]: DEBUG nova.policy [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fd6ed09d33f6491b83dc72cef31bfd29', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '812e115cc1114d0e8536924c4f5556f8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 673.559482] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "c0be3a6b-4d9d-4841-b05c-6aca4662b004" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.559635] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "c0be3a6b-4d9d-4841-b05c-6aca4662b004" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.609282] env[61868]: DEBUG nova.network.neutron [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Successfully created port: ff91543d-3c5c-4222-8955-f78dcfa1da3a {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 673.670107] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45d2078d-b5fa-4e29-b947-f9f07449e078 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.678369] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97a0395b-5e85-4d76-bdc4-8c8b68a09bd8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.710009] env[61868]: DEBUG nova.compute.manager [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 673.715081] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a75e69b3-e954-4d23-ab22-ef7d8aaad38a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.722736] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c87be83a-7c5f-4431-b683-c1ca3a492963 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.740160] env[61868]: DEBUG oslo_concurrency.lockutils [None req-74258ca0-a7fd-4b19-807a-dd66b3b700ec tempest-ServersTestFqdnHostnames-1266946302 tempest-ServersTestFqdnHostnames-1266946302-project-member] Lock "1586080c-3406-47f0-bbb9-0c50082eaf8d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 100.809s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.744037] env[61868]: DEBUG nova.compute.provider_tree [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 674.249331] env[61868]: DEBUG nova.scheduler.client.report [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 674.255012] env[61868]: DEBUG nova.compute.manager [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 674.349414] env[61868]: DEBUG nova.compute.manager [req-8d0d3d4f-d8c1-40c8-8e55-34435f518a72 req-19536ce8-2162-46d6-be6a-eccc918345ed service nova] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Received event network-changed-ff91543d-3c5c-4222-8955-f78dcfa1da3a {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 674.349414] env[61868]: DEBUG nova.compute.manager [req-8d0d3d4f-d8c1-40c8-8e55-34435f518a72 req-19536ce8-2162-46d6-be6a-eccc918345ed service nova] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Refreshing instance network info cache due to event network-changed-ff91543d-3c5c-4222-8955-f78dcfa1da3a. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 674.349414] env[61868]: DEBUG oslo_concurrency.lockutils [req-8d0d3d4f-d8c1-40c8-8e55-34435f518a72 req-19536ce8-2162-46d6-be6a-eccc918345ed service nova] Acquiring lock "refresh_cache-0d80665a-c881-454a-90c7-96348e7fab74" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 674.349414] env[61868]: DEBUG oslo_concurrency.lockutils [req-8d0d3d4f-d8c1-40c8-8e55-34435f518a72 req-19536ce8-2162-46d6-be6a-eccc918345ed service nova] Acquired lock "refresh_cache-0d80665a-c881-454a-90c7-96348e7fab74" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.349414] env[61868]: DEBUG nova.network.neutron [req-8d0d3d4f-d8c1-40c8-8e55-34435f518a72 req-19536ce8-2162-46d6-be6a-eccc918345ed service nova] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Refreshing network info cache for port ff91543d-3c5c-4222-8955-f78dcfa1da3a {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 674.696490] env[61868]: ERROR nova.compute.manager [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ff91543d-3c5c-4222-8955-f78dcfa1da3a, please check neutron logs for more information. [ 674.696490] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 674.696490] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 674.696490] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 674.696490] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 674.696490] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 674.696490] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 674.696490] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 674.696490] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 674.696490] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 674.696490] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 674.696490] env[61868]: ERROR nova.compute.manager raise self.value [ 674.696490] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 674.696490] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 674.696490] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 674.696490] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 674.696963] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 674.696963] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 674.696963] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ff91543d-3c5c-4222-8955-f78dcfa1da3a, please check neutron logs for more information. [ 674.696963] env[61868]: ERROR nova.compute.manager [ 674.696963] env[61868]: Traceback (most recent call last): [ 674.696963] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 674.696963] env[61868]: listener.cb(fileno) [ 674.696963] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 674.696963] env[61868]: result = function(*args, **kwargs) [ 674.696963] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 674.696963] env[61868]: return func(*args, **kwargs) [ 674.696963] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 674.696963] env[61868]: raise e [ 674.696963] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 674.696963] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 674.696963] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 674.696963] env[61868]: created_port_ids = self._update_ports_for_instance( [ 674.696963] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 674.696963] env[61868]: with excutils.save_and_reraise_exception(): [ 674.696963] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 674.696963] env[61868]: self.force_reraise() [ 674.696963] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 674.696963] env[61868]: raise self.value [ 674.696963] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 674.696963] env[61868]: updated_port = self._update_port( [ 674.696963] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 674.696963] env[61868]: _ensure_no_port_binding_failure(port) [ 674.696963] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 674.696963] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 674.697773] env[61868]: nova.exception.PortBindingFailed: Binding failed for port ff91543d-3c5c-4222-8955-f78dcfa1da3a, please check neutron logs for more information. [ 674.697773] env[61868]: Removing descriptor: 15 [ 674.721289] env[61868]: DEBUG nova.compute.manager [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 674.758351] env[61868]: DEBUG nova.virt.hardware [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 674.758792] env[61868]: DEBUG nova.virt.hardware [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 674.759343] env[61868]: DEBUG nova.virt.hardware [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 674.762862] env[61868]: DEBUG nova.virt.hardware [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 674.762862] env[61868]: DEBUG nova.virt.hardware [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 674.762862] env[61868]: DEBUG nova.virt.hardware [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 674.762862] env[61868]: DEBUG nova.virt.hardware [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 674.762862] env[61868]: DEBUG nova.virt.hardware [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 674.763203] env[61868]: DEBUG nova.virt.hardware [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 674.763203] env[61868]: DEBUG nova.virt.hardware [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 674.763203] env[61868]: DEBUG nova.virt.hardware [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 674.763203] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.071s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.763203] env[61868]: ERROR nova.compute.manager [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0aeb4b03-3fbd-4eb0-9261-fb6dc4b6077e, please check neutron logs for more information. [ 674.763355] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Traceback (most recent call last): [ 674.763355] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 674.763355] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] self.driver.spawn(context, instance, image_meta, [ 674.763355] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 674.763355] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] self._vmops.spawn(context, instance, image_meta, injected_files, [ 674.763355] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 674.763355] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] vm_ref = self.build_virtual_machine(instance, [ 674.763355] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 674.763355] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] vif_infos = vmwarevif.get_vif_info(self._session, [ 674.763355] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 674.763355] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] for vif in network_info: [ 674.763355] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 674.763355] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] return self._sync_wrapper(fn, *args, **kwargs) [ 674.763705] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 674.763705] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] self.wait() [ 674.763705] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 674.763705] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] self[:] = self._gt.wait() [ 674.763705] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 674.763705] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] return self._exit_event.wait() [ 674.763705] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 674.763705] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] result = hub.switch() [ 674.763705] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 674.763705] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] return self.greenlet.switch() [ 674.763705] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 674.763705] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] result = function(*args, **kwargs) [ 674.763705] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 674.764023] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] return func(*args, **kwargs) [ 674.764023] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 674.764023] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] raise e [ 674.764023] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 674.764023] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] nwinfo = self.network_api.allocate_for_instance( [ 674.764023] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 674.764023] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] created_port_ids = self._update_ports_for_instance( [ 674.764023] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 674.764023] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] with excutils.save_and_reraise_exception(): [ 674.764023] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 674.764023] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] self.force_reraise() [ 674.764023] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 674.764023] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] raise self.value [ 674.764330] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 674.764330] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] updated_port = self._update_port( [ 674.764330] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 674.764330] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] _ensure_no_port_binding_failure(port) [ 674.764330] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 674.764330] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] raise exception.PortBindingFailed(port_id=port['id']) [ 674.764330] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] nova.exception.PortBindingFailed: Binding failed for port 0aeb4b03-3fbd-4eb0-9261-fb6dc4b6077e, please check neutron logs for more information. [ 674.764330] env[61868]: ERROR nova.compute.manager [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] [ 674.764330] env[61868]: DEBUG nova.compute.utils [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Binding failed for port 0aeb4b03-3fbd-4eb0-9261-fb6dc4b6077e, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 674.767062] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e980b91-e30c-41fa-94dd-080de9bfabd3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.772153] env[61868]: DEBUG nova.compute.manager [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Build of instance ae5ec212-9dbe-4a11-94a3-8aa4576ca4db was re-scheduled: Binding failed for port 0aeb4b03-3fbd-4eb0-9261-fb6dc4b6077e, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 674.772911] env[61868]: DEBUG nova.compute.manager [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 674.773021] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] Acquiring lock "refresh_cache-ae5ec212-9dbe-4a11-94a3-8aa4576ca4db" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 674.773418] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] Acquired lock "refresh_cache-ae5ec212-9dbe-4a11-94a3-8aa4576ca4db" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.773786] env[61868]: DEBUG nova.network.neutron [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 674.774916] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.599s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.777100] env[61868]: INFO nova.compute.claims [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 674.786640] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11976bb8-d966-4550-b587-53c8fa7e90c1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.808815] env[61868]: ERROR nova.compute.manager [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ff91543d-3c5c-4222-8955-f78dcfa1da3a, please check neutron logs for more information. [ 674.808815] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Traceback (most recent call last): [ 674.808815] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 674.808815] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] yield resources [ 674.808815] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 674.808815] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] self.driver.spawn(context, instance, image_meta, [ 674.808815] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 674.808815] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] self._vmops.spawn(context, instance, image_meta, injected_files, [ 674.808815] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 674.808815] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] vm_ref = self.build_virtual_machine(instance, [ 674.808815] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 674.810073] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] vif_infos = vmwarevif.get_vif_info(self._session, [ 674.810073] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 674.810073] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] for vif in network_info: [ 674.810073] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 674.810073] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] return self._sync_wrapper(fn, *args, **kwargs) [ 674.810073] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 674.810073] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] self.wait() [ 674.810073] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 674.810073] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] self[:] = self._gt.wait() [ 674.810073] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 674.810073] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] return self._exit_event.wait() [ 674.810073] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 674.810073] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] current.throw(*self._exc) [ 674.810463] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 674.810463] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] result = function(*args, **kwargs) [ 674.810463] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 674.810463] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] return func(*args, **kwargs) [ 674.810463] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 674.810463] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] raise e [ 674.810463] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 674.810463] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] nwinfo = self.network_api.allocate_for_instance( [ 674.810463] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 674.810463] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] created_port_ids = self._update_ports_for_instance( [ 674.810463] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 674.810463] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] with excutils.save_and_reraise_exception(): [ 674.810463] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 674.810781] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] self.force_reraise() [ 674.810781] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 674.810781] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] raise self.value [ 674.810781] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 674.810781] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] updated_port = self._update_port( [ 674.810781] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 674.810781] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] _ensure_no_port_binding_failure(port) [ 674.810781] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 674.810781] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] raise exception.PortBindingFailed(port_id=port['id']) [ 674.810781] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] nova.exception.PortBindingFailed: Binding failed for port ff91543d-3c5c-4222-8955-f78dcfa1da3a, please check neutron logs for more information. [ 674.810781] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] [ 674.810781] env[61868]: INFO nova.compute.manager [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Terminating instance [ 674.812403] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.813304] env[61868]: DEBUG oslo_concurrency.lockutils [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "refresh_cache-0d80665a-c881-454a-90c7-96348e7fab74" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 674.866659] env[61868]: DEBUG nova.network.neutron [req-8d0d3d4f-d8c1-40c8-8e55-34435f518a72 req-19536ce8-2162-46d6-be6a-eccc918345ed service nova] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 674.946015] env[61868]: DEBUG nova.network.neutron [req-8d0d3d4f-d8c1-40c8-8e55-34435f518a72 req-19536ce8-2162-46d6-be6a-eccc918345ed service nova] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 675.305845] env[61868]: DEBUG nova.network.neutron [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 675.435394] env[61868]: DEBUG nova.network.neutron [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 675.449097] env[61868]: DEBUG oslo_concurrency.lockutils [req-8d0d3d4f-d8c1-40c8-8e55-34435f518a72 req-19536ce8-2162-46d6-be6a-eccc918345ed service nova] Releasing lock "refresh_cache-0d80665a-c881-454a-90c7-96348e7fab74" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 675.449532] env[61868]: DEBUG oslo_concurrency.lockutils [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquired lock "refresh_cache-0d80665a-c881-454a-90c7-96348e7fab74" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.449716] env[61868]: DEBUG nova.network.neutron [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 675.943247] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] Releasing lock "refresh_cache-ae5ec212-9dbe-4a11-94a3-8aa4576ca4db" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 675.943410] env[61868]: DEBUG nova.compute.manager [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 675.943598] env[61868]: DEBUG nova.compute.manager [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 675.943765] env[61868]: DEBUG nova.network.neutron [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 675.981496] env[61868]: DEBUG nova.network.neutron [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 675.984386] env[61868]: DEBUG nova.network.neutron [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 676.075368] env[61868]: DEBUG nova.network.neutron [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.156490] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fce24b38-04ca-4a19-9329-369fecb4535c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.165308] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6ee80e1-fbaf-4ecb-bb98-1ae1a5c9efd2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.195364] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e62eea1b-06e1-4cdf-b82a-720cc292a097 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.202154] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5b488f1-3c7f-4c7f-8ca2-0ffc6372bbd6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.216430] env[61868]: DEBUG nova.compute.provider_tree [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 676.373153] env[61868]: DEBUG nova.compute.manager [req-03ec44b8-77e5-4925-b6b4-f63fa1d0cfa6 req-f7308d68-0248-42d0-addc-bae4f09f516a service nova] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Received event network-vif-deleted-ff91543d-3c5c-4222-8955-f78dcfa1da3a {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 676.484598] env[61868]: DEBUG nova.network.neutron [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.580071] env[61868]: DEBUG oslo_concurrency.lockutils [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Releasing lock "refresh_cache-0d80665a-c881-454a-90c7-96348e7fab74" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 676.583053] env[61868]: DEBUG nova.compute.manager [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 676.583053] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 676.583053] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2e0b7a09-eed4-46ce-93f1-95fe399e144e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.592441] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aa30a08-b090-4e59-9828-300cc9de8ca2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.612406] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0d80665a-c881-454a-90c7-96348e7fab74 could not be found. [ 676.612649] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 676.612826] env[61868]: INFO nova.compute.manager [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Took 0.03 seconds to destroy the instance on the hypervisor. [ 676.613085] env[61868]: DEBUG oslo.service.loopingcall [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 676.613313] env[61868]: DEBUG nova.compute.manager [-] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 676.613410] env[61868]: DEBUG nova.network.neutron [-] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 676.629081] env[61868]: DEBUG nova.network.neutron [-] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 676.719135] env[61868]: DEBUG nova.scheduler.client.report [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 676.986986] env[61868]: INFO nova.compute.manager [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] [instance: ae5ec212-9dbe-4a11-94a3-8aa4576ca4db] Took 1.04 seconds to deallocate network for instance. [ 677.131490] env[61868]: DEBUG nova.network.neutron [-] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.224995] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.450s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.225805] env[61868]: DEBUG nova.compute.manager [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 677.228662] env[61868]: DEBUG oslo_concurrency.lockutils [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.130s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 677.634233] env[61868]: INFO nova.compute.manager [-] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Took 1.02 seconds to deallocate network for instance. [ 677.636590] env[61868]: DEBUG nova.compute.claims [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 677.636763] env[61868]: DEBUG oslo_concurrency.lockutils [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 677.734387] env[61868]: DEBUG nova.compute.utils [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 677.738622] env[61868]: DEBUG nova.compute.manager [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 677.740267] env[61868]: DEBUG nova.network.neutron [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 677.789653] env[61868]: DEBUG nova.policy [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '03ead7f08e934c4e9a1b39fbe08ed6e3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2cb049f497b14ac39a0316132f5802fc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 678.016847] env[61868]: INFO nova.scheduler.client.report [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] Deleted allocations for instance ae5ec212-9dbe-4a11-94a3-8aa4576ca4db [ 678.050497] env[61868]: DEBUG nova.network.neutron [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Successfully created port: 6a100ab0-5a39-4a17-9ed1-87a211f37c0b {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 678.095401] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b46b44d5-a033-4c11-8713-8efb1499a822 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.103596] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e8f5d9e-9476-4342-b0dd-fee83d797408 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.134372] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d6c7b19-1333-49df-ba26-1413f1d4ad56 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.142118] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9bb97c9-c07a-4057-8e17-c2b4aa16b6d3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.156433] env[61868]: DEBUG nova.compute.provider_tree [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 678.239327] env[61868]: DEBUG nova.compute.manager [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 678.527543] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5124bc85-c80b-42fe-8ff7-21444f807676 tempest-ServerAddressesTestJSON-105959644 tempest-ServerAddressesTestJSON-105959644-project-member] Lock "ae5ec212-9dbe-4a11-94a3-8aa4576ca4db" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 101.930s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.664109] env[61868]: DEBUG nova.scheduler.client.report [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 678.886508] env[61868]: DEBUG nova.compute.manager [req-fd2130bc-0b29-4c38-b191-d21c6531818f req-c44f3537-fb75-44f5-a7d6-b7cdc07e2cc1 service nova] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Received event network-changed-6a100ab0-5a39-4a17-9ed1-87a211f37c0b {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 678.886775] env[61868]: DEBUG nova.compute.manager [req-fd2130bc-0b29-4c38-b191-d21c6531818f req-c44f3537-fb75-44f5-a7d6-b7cdc07e2cc1 service nova] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Refreshing instance network info cache due to event network-changed-6a100ab0-5a39-4a17-9ed1-87a211f37c0b. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 678.887902] env[61868]: DEBUG oslo_concurrency.lockutils [req-fd2130bc-0b29-4c38-b191-d21c6531818f req-c44f3537-fb75-44f5-a7d6-b7cdc07e2cc1 service nova] Acquiring lock "refresh_cache-b6118327-0376-4b05-aee3-cb56d9cdf0b2" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 678.887902] env[61868]: DEBUG oslo_concurrency.lockutils [req-fd2130bc-0b29-4c38-b191-d21c6531818f req-c44f3537-fb75-44f5-a7d6-b7cdc07e2cc1 service nova] Acquired lock "refresh_cache-b6118327-0376-4b05-aee3-cb56d9cdf0b2" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.887902] env[61868]: DEBUG nova.network.neutron [req-fd2130bc-0b29-4c38-b191-d21c6531818f req-c44f3537-fb75-44f5-a7d6-b7cdc07e2cc1 service nova] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Refreshing network info cache for port 6a100ab0-5a39-4a17-9ed1-87a211f37c0b {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 679.011081] env[61868]: ERROR nova.compute.manager [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6a100ab0-5a39-4a17-9ed1-87a211f37c0b, please check neutron logs for more information. [ 679.011081] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 679.011081] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 679.011081] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 679.011081] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 679.011081] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 679.011081] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 679.011081] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 679.011081] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 679.011081] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 679.011081] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 679.011081] env[61868]: ERROR nova.compute.manager raise self.value [ 679.011081] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 679.011081] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 679.011081] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 679.011081] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 679.011560] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 679.011560] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 679.011560] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6a100ab0-5a39-4a17-9ed1-87a211f37c0b, please check neutron logs for more information. [ 679.011560] env[61868]: ERROR nova.compute.manager [ 679.011560] env[61868]: Traceback (most recent call last): [ 679.011560] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 679.011560] env[61868]: listener.cb(fileno) [ 679.011560] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 679.011560] env[61868]: result = function(*args, **kwargs) [ 679.011560] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 679.011560] env[61868]: return func(*args, **kwargs) [ 679.011560] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 679.011560] env[61868]: raise e [ 679.011560] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 679.011560] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 679.011560] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 679.011560] env[61868]: created_port_ids = self._update_ports_for_instance( [ 679.011560] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 679.011560] env[61868]: with excutils.save_and_reraise_exception(): [ 679.011560] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 679.011560] env[61868]: self.force_reraise() [ 679.011560] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 679.011560] env[61868]: raise self.value [ 679.011560] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 679.011560] env[61868]: updated_port = self._update_port( [ 679.011560] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 679.011560] env[61868]: _ensure_no_port_binding_failure(port) [ 679.011560] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 679.011560] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 679.012402] env[61868]: nova.exception.PortBindingFailed: Binding failed for port 6a100ab0-5a39-4a17-9ed1-87a211f37c0b, please check neutron logs for more information. [ 679.012402] env[61868]: Removing descriptor: 16 [ 679.031112] env[61868]: DEBUG nova.compute.manager [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 679.168013] env[61868]: DEBUG oslo_concurrency.lockutils [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.939s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.168760] env[61868]: ERROR nova.compute.manager [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 54acfe88-357f-46cb-b86d-5fdf61ac8f40, please check neutron logs for more information. [ 679.168760] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Traceback (most recent call last): [ 679.168760] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 679.168760] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] self.driver.spawn(context, instance, image_meta, [ 679.168760] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 679.168760] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] self._vmops.spawn(context, instance, image_meta, injected_files, [ 679.168760] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 679.168760] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] vm_ref = self.build_virtual_machine(instance, [ 679.168760] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 679.168760] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] vif_infos = vmwarevif.get_vif_info(self._session, [ 679.168760] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 679.169051] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] for vif in network_info: [ 679.169051] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 679.169051] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] return self._sync_wrapper(fn, *args, **kwargs) [ 679.169051] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 679.169051] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] self.wait() [ 679.169051] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 679.169051] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] self[:] = self._gt.wait() [ 679.169051] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 679.169051] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] return self._exit_event.wait() [ 679.169051] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 679.169051] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] current.throw(*self._exc) [ 679.169051] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 679.169051] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] result = function(*args, **kwargs) [ 679.169372] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 679.169372] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] return func(*args, **kwargs) [ 679.169372] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 679.169372] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] raise e [ 679.169372] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 679.169372] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] nwinfo = self.network_api.allocate_for_instance( [ 679.169372] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 679.169372] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] created_port_ids = self._update_ports_for_instance( [ 679.169372] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 679.169372] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] with excutils.save_and_reraise_exception(): [ 679.169372] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 679.169372] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] self.force_reraise() [ 679.169372] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 679.169746] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] raise self.value [ 679.169746] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 679.169746] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] updated_port = self._update_port( [ 679.169746] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 679.169746] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] _ensure_no_port_binding_failure(port) [ 679.169746] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 679.169746] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] raise exception.PortBindingFailed(port_id=port['id']) [ 679.169746] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] nova.exception.PortBindingFailed: Binding failed for port 54acfe88-357f-46cb-b86d-5fdf61ac8f40, please check neutron logs for more information. [ 679.169746] env[61868]: ERROR nova.compute.manager [instance: a8105842-1649-495d-b0c4-18ceda62bd57] [ 679.169746] env[61868]: DEBUG nova.compute.utils [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Binding failed for port 54acfe88-357f-46cb-b86d-5fdf61ac8f40, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 679.170834] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.322s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.174460] env[61868]: DEBUG nova.compute.manager [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Build of instance a8105842-1649-495d-b0c4-18ceda62bd57 was re-scheduled: Binding failed for port 54acfe88-357f-46cb-b86d-5fdf61ac8f40, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 679.174893] env[61868]: DEBUG nova.compute.manager [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 679.175280] env[61868]: DEBUG oslo_concurrency.lockutils [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Acquiring lock "refresh_cache-a8105842-1649-495d-b0c4-18ceda62bd57" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 679.175440] env[61868]: DEBUG oslo_concurrency.lockutils [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Acquired lock "refresh_cache-a8105842-1649-495d-b0c4-18ceda62bd57" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.175606] env[61868]: DEBUG nova.network.neutron [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 679.249368] env[61868]: DEBUG nova.compute.manager [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 679.275578] env[61868]: DEBUG nova.virt.hardware [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 679.275978] env[61868]: DEBUG nova.virt.hardware [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 679.276262] env[61868]: DEBUG nova.virt.hardware [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 679.276499] env[61868]: DEBUG nova.virt.hardware [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 679.276775] env[61868]: DEBUG nova.virt.hardware [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 679.276859] env[61868]: DEBUG nova.virt.hardware [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 679.277028] env[61868]: DEBUG nova.virt.hardware [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 679.277200] env[61868]: DEBUG nova.virt.hardware [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 679.277366] env[61868]: DEBUG nova.virt.hardware [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 679.277532] env[61868]: DEBUG nova.virt.hardware [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 679.277702] env[61868]: DEBUG nova.virt.hardware [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 679.278876] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1946e2b5-536a-42a0-a6c0-f3a5236df4d5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.287162] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a05ed037-cc2e-4e0c-89da-fefcf9921860 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.301921] env[61868]: ERROR nova.compute.manager [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6a100ab0-5a39-4a17-9ed1-87a211f37c0b, please check neutron logs for more information. [ 679.301921] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Traceback (most recent call last): [ 679.301921] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 679.301921] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] yield resources [ 679.301921] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 679.301921] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] self.driver.spawn(context, instance, image_meta, [ 679.301921] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 679.301921] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 679.301921] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 679.301921] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] vm_ref = self.build_virtual_machine(instance, [ 679.301921] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 679.302379] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] vif_infos = vmwarevif.get_vif_info(self._session, [ 679.302379] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 679.302379] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] for vif in network_info: [ 679.302379] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 679.302379] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] return self._sync_wrapper(fn, *args, **kwargs) [ 679.302379] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 679.302379] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] self.wait() [ 679.302379] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 679.302379] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] self[:] = self._gt.wait() [ 679.302379] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 679.302379] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] return self._exit_event.wait() [ 679.302379] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 679.302379] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] current.throw(*self._exc) [ 679.302961] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 679.302961] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] result = function(*args, **kwargs) [ 679.302961] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 679.302961] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] return func(*args, **kwargs) [ 679.302961] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 679.302961] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] raise e [ 679.302961] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 679.302961] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] nwinfo = self.network_api.allocate_for_instance( [ 679.302961] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 679.302961] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] created_port_ids = self._update_ports_for_instance( [ 679.302961] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 679.302961] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] with excutils.save_and_reraise_exception(): [ 679.302961] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 679.303662] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] self.force_reraise() [ 679.303662] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 679.303662] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] raise self.value [ 679.303662] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 679.303662] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] updated_port = self._update_port( [ 679.303662] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 679.303662] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] _ensure_no_port_binding_failure(port) [ 679.303662] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 679.303662] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] raise exception.PortBindingFailed(port_id=port['id']) [ 679.303662] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] nova.exception.PortBindingFailed: Binding failed for port 6a100ab0-5a39-4a17-9ed1-87a211f37c0b, please check neutron logs for more information. [ 679.303662] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] [ 679.303662] env[61868]: INFO nova.compute.manager [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Terminating instance [ 679.304476] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquiring lock "refresh_cache-b6118327-0376-4b05-aee3-cb56d9cdf0b2" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 679.406487] env[61868]: DEBUG nova.network.neutron [req-fd2130bc-0b29-4c38-b191-d21c6531818f req-c44f3537-fb75-44f5-a7d6-b7cdc07e2cc1 service nova] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 679.494383] env[61868]: DEBUG nova.network.neutron [req-fd2130bc-0b29-4c38-b191-d21c6531818f req-c44f3537-fb75-44f5-a7d6-b7cdc07e2cc1 service nova] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.553624] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 679.703238] env[61868]: DEBUG nova.network.neutron [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 679.752309] env[61868]: DEBUG nova.network.neutron [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.999299] env[61868]: DEBUG oslo_concurrency.lockutils [req-fd2130bc-0b29-4c38-b191-d21c6531818f req-c44f3537-fb75-44f5-a7d6-b7cdc07e2cc1 service nova] Releasing lock "refresh_cache-b6118327-0376-4b05-aee3-cb56d9cdf0b2" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 679.999944] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquired lock "refresh_cache-b6118327-0376-4b05-aee3-cb56d9cdf0b2" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.999944] env[61868]: DEBUG nova.network.neutron [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 680.012722] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14317c05-ebe3-4982-871c-d3ada715389e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.021727] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-429118be-fde8-4ae7-95a0-e33ac6d0b1aa {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.052793] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82bf5d06-ff5d-43c7-9cf9-107b6859083f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.060771] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57ae8c3f-ff30-41fa-8b61-814f27b8b72d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.079970] env[61868]: DEBUG nova.compute.provider_tree [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 680.254792] env[61868]: DEBUG oslo_concurrency.lockutils [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Releasing lock "refresh_cache-a8105842-1649-495d-b0c4-18ceda62bd57" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 680.255080] env[61868]: DEBUG nova.compute.manager [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 680.255201] env[61868]: DEBUG nova.compute.manager [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 680.255366] env[61868]: DEBUG nova.network.neutron [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 680.272878] env[61868]: DEBUG nova.network.neutron [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 680.517720] env[61868]: DEBUG nova.network.neutron [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 680.586206] env[61868]: DEBUG nova.scheduler.client.report [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 680.604656] env[61868]: DEBUG nova.network.neutron [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.776164] env[61868]: DEBUG nova.network.neutron [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.935040] env[61868]: DEBUG nova.compute.manager [req-c47d1fc3-3d45-4169-8446-1aac8811c8ef req-9286bf61-cd06-4756-8fb1-851a1bed17f5 service nova] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Received event network-vif-deleted-6a100ab0-5a39-4a17-9ed1-87a211f37c0b {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 681.092701] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.922s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.093337] env[61868]: ERROR nova.compute.manager [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 59d1a6f1-3a24-4ce3-b41d-442fff832e84, please check neutron logs for more information. [ 681.093337] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Traceback (most recent call last): [ 681.093337] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 681.093337] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] self.driver.spawn(context, instance, image_meta, [ 681.093337] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 681.093337] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] self._vmops.spawn(context, instance, image_meta, injected_files, [ 681.093337] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 681.093337] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] vm_ref = self.build_virtual_machine(instance, [ 681.093337] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 681.093337] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] vif_infos = vmwarevif.get_vif_info(self._session, [ 681.093337] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 681.093643] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] for vif in network_info: [ 681.093643] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 681.093643] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] return self._sync_wrapper(fn, *args, **kwargs) [ 681.093643] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 681.093643] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] self.wait() [ 681.093643] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 681.093643] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] self[:] = self._gt.wait() [ 681.093643] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 681.093643] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] return self._exit_event.wait() [ 681.093643] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 681.093643] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] current.throw(*self._exc) [ 681.093643] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 681.093643] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] result = function(*args, **kwargs) [ 681.093949] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 681.093949] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] return func(*args, **kwargs) [ 681.093949] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 681.093949] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] raise e [ 681.093949] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 681.093949] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] nwinfo = self.network_api.allocate_for_instance( [ 681.093949] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 681.093949] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] created_port_ids = self._update_ports_for_instance( [ 681.093949] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 681.093949] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] with excutils.save_and_reraise_exception(): [ 681.093949] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 681.093949] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] self.force_reraise() [ 681.093949] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 681.094278] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] raise self.value [ 681.094278] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 681.094278] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] updated_port = self._update_port( [ 681.094278] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 681.094278] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] _ensure_no_port_binding_failure(port) [ 681.094278] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 681.094278] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] raise exception.PortBindingFailed(port_id=port['id']) [ 681.094278] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] nova.exception.PortBindingFailed: Binding failed for port 59d1a6f1-3a24-4ce3-b41d-442fff832e84, please check neutron logs for more information. [ 681.094278] env[61868]: ERROR nova.compute.manager [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] [ 681.094278] env[61868]: DEBUG nova.compute.utils [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Binding failed for port 59d1a6f1-3a24-4ce3-b41d-442fff832e84, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 681.095224] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.719s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 681.098104] env[61868]: DEBUG nova.compute.manager [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Build of instance 2d1b59b1-504c-487f-93ac-642c490d6546 was re-scheduled: Binding failed for port 59d1a6f1-3a24-4ce3-b41d-442fff832e84, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 681.098535] env[61868]: DEBUG nova.compute.manager [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 681.098758] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] Acquiring lock "refresh_cache-2d1b59b1-504c-487f-93ac-642c490d6546" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 681.098901] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] Acquired lock "refresh_cache-2d1b59b1-504c-487f-93ac-642c490d6546" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.099070] env[61868]: DEBUG nova.network.neutron [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 681.107068] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Releasing lock "refresh_cache-b6118327-0376-4b05-aee3-cb56d9cdf0b2" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.107449] env[61868]: DEBUG nova.compute.manager [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 681.107638] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 681.107906] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c8f13217-f738-4e23-a6fe-d192ddfbc077 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.117754] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1052bc35-2a32-4e94-a1e4-233705b0dbe9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.141812] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b6118327-0376-4b05-aee3-cb56d9cdf0b2 could not be found. [ 681.142228] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 681.142228] env[61868]: INFO nova.compute.manager [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Took 0.03 seconds to destroy the instance on the hypervisor. [ 681.142468] env[61868]: DEBUG oslo.service.loopingcall [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 681.142930] env[61868]: DEBUG nova.compute.manager [-] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 681.143034] env[61868]: DEBUG nova.network.neutron [-] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 681.158993] env[61868]: DEBUG nova.network.neutron [-] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 681.278298] env[61868]: INFO nova.compute.manager [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] [instance: a8105842-1649-495d-b0c4-18ceda62bd57] Took 1.02 seconds to deallocate network for instance. [ 681.629521] env[61868]: DEBUG nova.network.neutron [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 681.660481] env[61868]: DEBUG nova.network.neutron [-] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.855338] env[61868]: DEBUG nova.network.neutron [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.004423] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7651f9c3-43b3-4358-8573-18746f4fdb2d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.011871] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cefdc3e5-fb09-4c6a-bd64-0c7dc956b9a4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.042566] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b500919c-32eb-44bc-8915-6927ad46e56b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.050784] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfd45eb0-e00e-415a-92d7-7e526c32360b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.064518] env[61868]: DEBUG nova.compute.provider_tree [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 682.163532] env[61868]: INFO nova.compute.manager [-] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Took 1.02 seconds to deallocate network for instance. [ 682.165902] env[61868]: DEBUG nova.compute.claims [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 682.166090] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 682.318373] env[61868]: INFO nova.scheduler.client.report [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Deleted allocations for instance a8105842-1649-495d-b0c4-18ceda62bd57 [ 682.358809] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] Releasing lock "refresh_cache-2d1b59b1-504c-487f-93ac-642c490d6546" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 682.359067] env[61868]: DEBUG nova.compute.manager [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 682.359257] env[61868]: DEBUG nova.compute.manager [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 682.359445] env[61868]: DEBUG nova.network.neutron [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 682.391620] env[61868]: DEBUG nova.network.neutron [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 682.568384] env[61868]: DEBUG nova.scheduler.client.report [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 682.826265] env[61868]: DEBUG oslo_concurrency.lockutils [None req-14bfc2d1-de40-4e4e-8a25-83305e524078 tempest-DeleteServersAdminTestJSON-1300319601 tempest-DeleteServersAdminTestJSON-1300319601-project-member] Lock "a8105842-1649-495d-b0c4-18ceda62bd57" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 104.687s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.894952] env[61868]: DEBUG nova.network.neutron [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.073611] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.978s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 683.074308] env[61868]: ERROR nova.compute.manager [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c7376ab9-07e7-4fe6-beac-650b9087fd7e, please check neutron logs for more information. [ 683.074308] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Traceback (most recent call last): [ 683.074308] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 683.074308] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] self.driver.spawn(context, instance, image_meta, [ 683.074308] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 683.074308] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 683.074308] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 683.074308] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] vm_ref = self.build_virtual_machine(instance, [ 683.074308] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 683.074308] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] vif_infos = vmwarevif.get_vif_info(self._session, [ 683.074308] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 683.074636] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] for vif in network_info: [ 683.074636] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 683.074636] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] return self._sync_wrapper(fn, *args, **kwargs) [ 683.074636] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 683.074636] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] self.wait() [ 683.074636] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 683.074636] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] self[:] = self._gt.wait() [ 683.074636] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 683.074636] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] return self._exit_event.wait() [ 683.074636] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 683.074636] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] current.throw(*self._exc) [ 683.074636] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 683.074636] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] result = function(*args, **kwargs) [ 683.074941] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 683.074941] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] return func(*args, **kwargs) [ 683.074941] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 683.074941] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] raise e [ 683.074941] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 683.074941] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] nwinfo = self.network_api.allocate_for_instance( [ 683.074941] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 683.074941] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] created_port_ids = self._update_ports_for_instance( [ 683.074941] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 683.074941] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] with excutils.save_and_reraise_exception(): [ 683.074941] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 683.074941] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] self.force_reraise() [ 683.074941] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 683.075268] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] raise self.value [ 683.075268] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 683.075268] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] updated_port = self._update_port( [ 683.075268] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 683.075268] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] _ensure_no_port_binding_failure(port) [ 683.075268] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 683.075268] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] raise exception.PortBindingFailed(port_id=port['id']) [ 683.075268] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] nova.exception.PortBindingFailed: Binding failed for port c7376ab9-07e7-4fe6-beac-650b9087fd7e, please check neutron logs for more information. [ 683.075268] env[61868]: ERROR nova.compute.manager [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] [ 683.075268] env[61868]: DEBUG nova.compute.utils [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Binding failed for port c7376ab9-07e7-4fe6-beac-650b9087fd7e, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 683.077113] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.269s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 683.081434] env[61868]: DEBUG nova.compute.manager [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Build of instance 3e970ecb-c842-4706-afd0-176d0088e1d3 was re-scheduled: Binding failed for port c7376ab9-07e7-4fe6-beac-650b9087fd7e, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 683.081871] env[61868]: DEBUG nova.compute.manager [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 683.082103] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Acquiring lock "refresh_cache-3e970ecb-c842-4706-afd0-176d0088e1d3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 683.082254] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Acquired lock "refresh_cache-3e970ecb-c842-4706-afd0-176d0088e1d3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.082414] env[61868]: DEBUG nova.network.neutron [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 683.329231] env[61868]: DEBUG nova.compute.manager [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 683.399365] env[61868]: INFO nova.compute.manager [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] [instance: 2d1b59b1-504c-487f-93ac-642c490d6546] Took 1.04 seconds to deallocate network for instance. [ 683.605912] env[61868]: DEBUG nova.network.neutron [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 683.692131] env[61868]: DEBUG nova.network.neutron [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.853573] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 683.996844] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29e2a07e-47e8-40b4-ae52-35613944d5b8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.004825] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f84d0c2-daaa-42ff-8a2a-b457afd92231 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.041050] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0124ef9-5553-4132-b078-62fbefc4d2a0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.051806] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a008493d-72bb-489d-ae6e-303f0b8943e6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.062316] env[61868]: DEBUG nova.compute.provider_tree [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 684.195790] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Releasing lock "refresh_cache-3e970ecb-c842-4706-afd0-176d0088e1d3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 684.195790] env[61868]: DEBUG nova.compute.manager [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 684.195790] env[61868]: DEBUG nova.compute.manager [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 684.196392] env[61868]: DEBUG nova.network.neutron [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 684.213858] env[61868]: DEBUG nova.network.neutron [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 684.453822] env[61868]: INFO nova.scheduler.client.report [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] Deleted allocations for instance 2d1b59b1-504c-487f-93ac-642c490d6546 [ 684.568170] env[61868]: DEBUG nova.scheduler.client.report [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 684.716379] env[61868]: DEBUG nova.network.neutron [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.964058] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6e63c4bf-1da7-4869-95b0-f70eeafb82ea tempest-ServerActionsTestJSON-1561567447 tempest-ServerActionsTestJSON-1561567447-project-member] Lock "2d1b59b1-504c-487f-93ac-642c490d6546" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 105.522s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.070616] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.994s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.071243] env[61868]: ERROR nova.compute.manager [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7c54b4ab-e267-4ab2-856a-5b7ac33984c7, please check neutron logs for more information. [ 685.071243] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Traceback (most recent call last): [ 685.071243] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 685.071243] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] self.driver.spawn(context, instance, image_meta, [ 685.071243] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 685.071243] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 685.071243] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 685.071243] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] vm_ref = self.build_virtual_machine(instance, [ 685.071243] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 685.071243] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] vif_infos = vmwarevif.get_vif_info(self._session, [ 685.071243] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 685.071721] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] for vif in network_info: [ 685.071721] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 685.071721] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] return self._sync_wrapper(fn, *args, **kwargs) [ 685.071721] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 685.071721] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] self.wait() [ 685.071721] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 685.071721] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] self[:] = self._gt.wait() [ 685.071721] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 685.071721] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] return self._exit_event.wait() [ 685.071721] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 685.071721] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] current.throw(*self._exc) [ 685.071721] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 685.071721] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] result = function(*args, **kwargs) [ 685.072102] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 685.072102] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] return func(*args, **kwargs) [ 685.072102] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 685.072102] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] raise e [ 685.072102] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 685.072102] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] nwinfo = self.network_api.allocate_for_instance( [ 685.072102] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 685.072102] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] created_port_ids = self._update_ports_for_instance( [ 685.072102] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 685.072102] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] with excutils.save_and_reraise_exception(): [ 685.072102] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 685.072102] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] self.force_reraise() [ 685.072102] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 685.072486] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] raise self.value [ 685.072486] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 685.072486] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] updated_port = self._update_port( [ 685.072486] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 685.072486] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] _ensure_no_port_binding_failure(port) [ 685.072486] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 685.072486] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] raise exception.PortBindingFailed(port_id=port['id']) [ 685.072486] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] nova.exception.PortBindingFailed: Binding failed for port 7c54b4ab-e267-4ab2-856a-5b7ac33984c7, please check neutron logs for more information. [ 685.072486] env[61868]: ERROR nova.compute.manager [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] [ 685.072486] env[61868]: DEBUG nova.compute.utils [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Binding failed for port 7c54b4ab-e267-4ab2-856a-5b7ac33984c7, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 685.073164] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 16.937s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.075425] env[61868]: DEBUG nova.compute.manager [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Build of instance 85dbaf67-049c-4a90-a3dc-67e2decc1dbf was re-scheduled: Binding failed for port 7c54b4ab-e267-4ab2-856a-5b7ac33984c7, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 685.075842] env[61868]: DEBUG nova.compute.manager [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 685.076150] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Acquiring lock "refresh_cache-85dbaf67-049c-4a90-a3dc-67e2decc1dbf" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 685.076214] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Acquired lock "refresh_cache-85dbaf67-049c-4a90-a3dc-67e2decc1dbf" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.076372] env[61868]: DEBUG nova.network.neutron [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 685.223022] env[61868]: INFO nova.compute.manager [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: 3e970ecb-c842-4706-afd0-176d0088e1d3] Took 1.02 seconds to deallocate network for instance. [ 685.410704] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Acquiring lock "cd1ccbd7-5d2e-4dd9-a62e-24706759a67a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 685.411299] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Lock "cd1ccbd7-5d2e-4dd9-a62e-24706759a67a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.468372] env[61868]: DEBUG nova.compute.manager [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 685.600404] env[61868]: DEBUG nova.network.neutron [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 685.725530] env[61868]: DEBUG nova.network.neutron [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.002221] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 686.237056] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Releasing lock "refresh_cache-85dbaf67-049c-4a90-a3dc-67e2decc1dbf" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 686.237056] env[61868]: DEBUG nova.compute.manager [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 686.237056] env[61868]: DEBUG nova.compute.manager [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 686.237056] env[61868]: DEBUG nova.network.neutron [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 686.258121] env[61868]: DEBUG nova.network.neutron [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 686.283137] env[61868]: INFO nova.scheduler.client.report [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Deleted allocations for instance 3e970ecb-c842-4706-afd0-176d0088e1d3 [ 686.605660] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 3e970ecb-c842-4706-afd0-176d0088e1d3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 686.764937] env[61868]: DEBUG nova.network.neutron [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.790948] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d6f2ca1-08af-43db-a6a3-911652d303d9 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Lock "3e970ecb-c842-4706-afd0-176d0088e1d3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 106.491s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.113360] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 85dbaf67-049c-4a90-a3dc-67e2decc1dbf has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 687.113558] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance c979d105-12b7-485c-a136-55ffad02fe67 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 687.113644] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 0d80665a-c881-454a-90c7-96348e7fab74 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 687.114403] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance b6118327-0376-4b05-aee3-cb56d9cdf0b2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 687.270121] env[61868]: INFO nova.compute.manager [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 85dbaf67-049c-4a90-a3dc-67e2decc1dbf] Took 1.03 seconds to deallocate network for instance. [ 687.296107] env[61868]: DEBUG nova.compute.manager [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 687.619343] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 96694bb9-fbf2-4b71-9cb5-37f81b634992 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 687.825118] env[61868]: DEBUG oslo_concurrency.lockutils [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 688.122986] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance c0c9e61b-ab2b-47dc-997b-570204ab4c3b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 688.300303] env[61868]: INFO nova.scheduler.client.report [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Deleted allocations for instance 85dbaf67-049c-4a90-a3dc-67e2decc1dbf [ 688.627049] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance cf7b7da2-380e-4a97-b142-0dd9e5ad3c79 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 688.787884] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] Acquiring lock "5c1683e1-0c87-4d1b-ae21-e07e04d848b1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 688.788187] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] Lock "5c1683e1-0c87-4d1b-ae21-e07e04d848b1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 688.808720] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6f71b97c-873a-4b46-ba37-2bc43f09866f tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Lock "85dbaf67-049c-4a90-a3dc-67e2decc1dbf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 107.504s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 689.130553] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 460255fb-6589-481b-9770-0a2b548c6194 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 689.311617] env[61868]: DEBUG nova.compute.manager [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 689.633868] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 689.840387] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 690.138226] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance b4a7ff58-0f48-4942-a270-de6910ff0c0c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 690.642242] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 92d3557c-655f-45d9-ae7d-bbe62567e1e2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 691.145754] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 66c5f446-91e5-4de1-abb6-f06a0a36020d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 691.650628] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 97d1d9c9-214f-4573-9efa-3254ea3f2ee0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 691.995796] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Acquiring lock "083140bf-233d-49ac-8247-cd93206f9200" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 691.997291] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Lock "083140bf-233d-49ac-8247-cd93206f9200" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 692.155168] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 84dc8092-c24a-4587-bfcb-865b6eb129aa has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 692.659832] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 693.162673] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 693.665951] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance df927c0c-b742-4451-9439-60ea63287b99 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 694.169207] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 4de32ee6-e205-435b-a75e-94c3121048e3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 694.673048] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance a0fda7cf-21b6-4a22-bf56-09e848b4dc9e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 695.176261] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance ebb09f50-50b8-49bb-b366-5b8906605699 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 695.679227] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 29530178-d69c-4aed-9061-d3d1cfa954d0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 696.182510] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance e023cfb8-5f1c-41d9-9e43-d10e073234ea has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 696.685586] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 6391addb-b1a5-4e15-9bd2-531eb9956b42 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 697.188587] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance c0be3a6b-4d9d-4841-b05c-6aca4662b004 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 697.692162] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance cd1ccbd7-5d2e-4dd9-a62e-24706759a67a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 697.692457] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=61868) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 697.693029] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=61868) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 697.984846] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-502aae7b-f58d-4c2e-97e5-a9ae456c3cde {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.992711] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3603586d-94af-4a4a-961b-a5705875c163 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.022113] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6d9e0ee-1440-4d54-b25d-3f17020ae835 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.029404] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03f06551-c100-40b9-a175-90b881522f9c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.042421] env[61868]: DEBUG nova.compute.provider_tree [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 698.546057] env[61868]: DEBUG nova.scheduler.client.report [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 699.051282] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61868) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 699.051282] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 13.978s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.051612] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.815s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 699.053147] env[61868]: INFO nova.compute.claims [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 700.333892] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c874be9-7206-4892-8cb7-13d0894fdf48 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.341218] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57b9b121-dfe9-4303-8fa2-be6de0df5bfe {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.369431] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b2ffbcd-7f36-4dcc-98aa-d7e6139f7408 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.376250] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13e6e4fc-b075-45fd-b2e7-acc3145e879a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.388478] env[61868]: DEBUG nova.compute.provider_tree [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 700.891921] env[61868]: DEBUG nova.scheduler.client.report [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 701.397049] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.345s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.397401] env[61868]: DEBUG nova.compute.manager [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 701.400162] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.157s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.905254] env[61868]: DEBUG nova.compute.utils [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 701.910812] env[61868]: DEBUG nova.compute.manager [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 701.910812] env[61868]: DEBUG nova.network.neutron [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 701.954358] env[61868]: DEBUG nova.policy [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c9f25c1f6cb74aa3a5353e11c8030c6a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c72bd3a3cc0b400fbb87d288ec5de723', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 702.257184] env[61868]: DEBUG nova.network.neutron [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Successfully created port: 48d4c453-40b0-4fb2-a83c-b145b3e20694 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 702.266808] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-714abcb5-d68d-45fc-ae97-b996a106e1cd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.277890] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02fd691d-1c91-4ec7-a0c5-866785463bac {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.310144] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15d14b3a-d068-45e4-812d-cebe5421e2cf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.318834] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-750bdc03-1a72-41e5-baf9-ae839d439da2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.335050] env[61868]: DEBUG nova.compute.provider_tree [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 702.411613] env[61868]: DEBUG nova.compute.manager [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 702.838191] env[61868]: DEBUG nova.scheduler.client.report [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 702.967190] env[61868]: DEBUG nova.compute.manager [req-7e8cc533-2b7b-4c18-ad70-8a3f54472a26 req-c56c75dc-bb98-4b0a-a224-9d86a6af917c service nova] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Received event network-changed-48d4c453-40b0-4fb2-a83c-b145b3e20694 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 702.967546] env[61868]: DEBUG nova.compute.manager [req-7e8cc533-2b7b-4c18-ad70-8a3f54472a26 req-c56c75dc-bb98-4b0a-a224-9d86a6af917c service nova] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Refreshing instance network info cache due to event network-changed-48d4c453-40b0-4fb2-a83c-b145b3e20694. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 702.967656] env[61868]: DEBUG oslo_concurrency.lockutils [req-7e8cc533-2b7b-4c18-ad70-8a3f54472a26 req-c56c75dc-bb98-4b0a-a224-9d86a6af917c service nova] Acquiring lock "refresh_cache-c0c9e61b-ab2b-47dc-997b-570204ab4c3b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 702.967822] env[61868]: DEBUG oslo_concurrency.lockutils [req-7e8cc533-2b7b-4c18-ad70-8a3f54472a26 req-c56c75dc-bb98-4b0a-a224-9d86a6af917c service nova] Acquired lock "refresh_cache-c0c9e61b-ab2b-47dc-997b-570204ab4c3b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.967990] env[61868]: DEBUG nova.network.neutron [req-7e8cc533-2b7b-4c18-ad70-8a3f54472a26 req-c56c75dc-bb98-4b0a-a224-9d86a6af917c service nova] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Refreshing network info cache for port 48d4c453-40b0-4fb2-a83c-b145b3e20694 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 703.151232] env[61868]: ERROR nova.compute.manager [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 48d4c453-40b0-4fb2-a83c-b145b3e20694, please check neutron logs for more information. [ 703.151232] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 703.151232] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 703.151232] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 703.151232] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 703.151232] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 703.151232] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 703.151232] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 703.151232] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 703.151232] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 703.151232] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 703.151232] env[61868]: ERROR nova.compute.manager raise self.value [ 703.151232] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 703.151232] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 703.151232] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 703.151232] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 703.151895] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 703.151895] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 703.151895] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 48d4c453-40b0-4fb2-a83c-b145b3e20694, please check neutron logs for more information. [ 703.151895] env[61868]: ERROR nova.compute.manager [ 703.151895] env[61868]: Traceback (most recent call last): [ 703.151895] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 703.151895] env[61868]: listener.cb(fileno) [ 703.151895] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 703.151895] env[61868]: result = function(*args, **kwargs) [ 703.151895] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 703.151895] env[61868]: return func(*args, **kwargs) [ 703.151895] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 703.151895] env[61868]: raise e [ 703.151895] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 703.151895] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 703.151895] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 703.151895] env[61868]: created_port_ids = self._update_ports_for_instance( [ 703.151895] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 703.151895] env[61868]: with excutils.save_and_reraise_exception(): [ 703.151895] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 703.151895] env[61868]: self.force_reraise() [ 703.151895] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 703.151895] env[61868]: raise self.value [ 703.151895] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 703.151895] env[61868]: updated_port = self._update_port( [ 703.151895] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 703.151895] env[61868]: _ensure_no_port_binding_failure(port) [ 703.151895] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 703.151895] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 703.153040] env[61868]: nova.exception.PortBindingFailed: Binding failed for port 48d4c453-40b0-4fb2-a83c-b145b3e20694, please check neutron logs for more information. [ 703.153040] env[61868]: Removing descriptor: 16 [ 703.344641] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.944s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.345297] env[61868]: ERROR nova.compute.manager [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ac679a13-bea7-4eeb-9517-68d1b5292e42, please check neutron logs for more information. [ 703.345297] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] Traceback (most recent call last): [ 703.345297] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 703.345297] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] self.driver.spawn(context, instance, image_meta, [ 703.345297] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 703.345297] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] self._vmops.spawn(context, instance, image_meta, injected_files, [ 703.345297] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 703.345297] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] vm_ref = self.build_virtual_machine(instance, [ 703.345297] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 703.345297] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] vif_infos = vmwarevif.get_vif_info(self._session, [ 703.345297] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 703.345601] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] for vif in network_info: [ 703.345601] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 703.345601] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] return self._sync_wrapper(fn, *args, **kwargs) [ 703.345601] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 703.345601] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] self.wait() [ 703.345601] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 703.345601] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] self[:] = self._gt.wait() [ 703.345601] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 703.345601] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] return self._exit_event.wait() [ 703.345601] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 703.345601] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] result = hub.switch() [ 703.345601] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 703.345601] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] return self.greenlet.switch() [ 703.345954] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 703.345954] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] result = function(*args, **kwargs) [ 703.345954] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 703.345954] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] return func(*args, **kwargs) [ 703.345954] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 703.345954] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] raise e [ 703.345954] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 703.345954] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] nwinfo = self.network_api.allocate_for_instance( [ 703.345954] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 703.345954] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] created_port_ids = self._update_ports_for_instance( [ 703.345954] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 703.345954] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] with excutils.save_and_reraise_exception(): [ 703.345954] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 703.346299] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] self.force_reraise() [ 703.346299] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 703.346299] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] raise self.value [ 703.346299] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 703.346299] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] updated_port = self._update_port( [ 703.346299] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 703.346299] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] _ensure_no_port_binding_failure(port) [ 703.346299] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 703.346299] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] raise exception.PortBindingFailed(port_id=port['id']) [ 703.346299] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] nova.exception.PortBindingFailed: Binding failed for port ac679a13-bea7-4eeb-9517-68d1b5292e42, please check neutron logs for more information. [ 703.346299] env[61868]: ERROR nova.compute.manager [instance: c979d105-12b7-485c-a136-55ffad02fe67] [ 703.346656] env[61868]: DEBUG nova.compute.utils [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Binding failed for port ac679a13-bea7-4eeb-9517-68d1b5292e42, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 703.347299] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.535s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.348783] env[61868]: INFO nova.compute.claims [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 703.351647] env[61868]: DEBUG nova.compute.manager [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Build of instance c979d105-12b7-485c-a136-55ffad02fe67 was re-scheduled: Binding failed for port ac679a13-bea7-4eeb-9517-68d1b5292e42, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 703.352431] env[61868]: DEBUG nova.compute.manager [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 703.352431] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Acquiring lock "refresh_cache-c979d105-12b7-485c-a136-55ffad02fe67" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 703.352540] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Acquired lock "refresh_cache-c979d105-12b7-485c-a136-55ffad02fe67" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.352604] env[61868]: DEBUG nova.network.neutron [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 703.422436] env[61868]: DEBUG nova.compute.manager [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 703.447509] env[61868]: DEBUG nova.virt.hardware [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 703.447752] env[61868]: DEBUG nova.virt.hardware [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 703.447906] env[61868]: DEBUG nova.virt.hardware [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 703.448099] env[61868]: DEBUG nova.virt.hardware [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 703.448248] env[61868]: DEBUG nova.virt.hardware [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 703.448393] env[61868]: DEBUG nova.virt.hardware [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 703.448725] env[61868]: DEBUG nova.virt.hardware [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 703.448953] env[61868]: DEBUG nova.virt.hardware [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 703.449204] env[61868]: DEBUG nova.virt.hardware [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 703.449381] env[61868]: DEBUG nova.virt.hardware [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 703.449557] env[61868]: DEBUG nova.virt.hardware [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 703.450404] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b56ccdc2-5d7d-4579-9ffe-2e96f34fd1cb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.458704] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3124438c-c0e2-4974-aa44-3f16971be564 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.475069] env[61868]: ERROR nova.compute.manager [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 48d4c453-40b0-4fb2-a83c-b145b3e20694, please check neutron logs for more information. [ 703.475069] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Traceback (most recent call last): [ 703.475069] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 703.475069] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] yield resources [ 703.475069] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 703.475069] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] self.driver.spawn(context, instance, image_meta, [ 703.475069] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 703.475069] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 703.475069] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 703.475069] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] vm_ref = self.build_virtual_machine(instance, [ 703.475069] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 703.475429] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] vif_infos = vmwarevif.get_vif_info(self._session, [ 703.475429] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 703.475429] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] for vif in network_info: [ 703.475429] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 703.475429] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] return self._sync_wrapper(fn, *args, **kwargs) [ 703.475429] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 703.475429] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] self.wait() [ 703.475429] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 703.475429] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] self[:] = self._gt.wait() [ 703.475429] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 703.475429] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] return self._exit_event.wait() [ 703.475429] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 703.475429] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] current.throw(*self._exc) [ 703.475755] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 703.475755] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] result = function(*args, **kwargs) [ 703.475755] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 703.475755] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] return func(*args, **kwargs) [ 703.475755] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 703.475755] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] raise e [ 703.475755] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 703.475755] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] nwinfo = self.network_api.allocate_for_instance( [ 703.475755] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 703.475755] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] created_port_ids = self._update_ports_for_instance( [ 703.475755] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 703.475755] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] with excutils.save_and_reraise_exception(): [ 703.475755] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 703.476087] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] self.force_reraise() [ 703.476087] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 703.476087] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] raise self.value [ 703.476087] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 703.476087] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] updated_port = self._update_port( [ 703.476087] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 703.476087] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] _ensure_no_port_binding_failure(port) [ 703.476087] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 703.476087] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] raise exception.PortBindingFailed(port_id=port['id']) [ 703.476087] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] nova.exception.PortBindingFailed: Binding failed for port 48d4c453-40b0-4fb2-a83c-b145b3e20694, please check neutron logs for more information. [ 703.476087] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] [ 703.476087] env[61868]: INFO nova.compute.manager [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Terminating instance [ 703.477467] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] Acquiring lock "refresh_cache-c0c9e61b-ab2b-47dc-997b-570204ab4c3b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 703.490057] env[61868]: DEBUG nova.network.neutron [req-7e8cc533-2b7b-4c18-ad70-8a3f54472a26 req-c56c75dc-bb98-4b0a-a224-9d86a6af917c service nova] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 703.571661] env[61868]: DEBUG nova.network.neutron [req-7e8cc533-2b7b-4c18-ad70-8a3f54472a26 req-c56c75dc-bb98-4b0a-a224-9d86a6af917c service nova] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.870947] env[61868]: DEBUG nova.network.neutron [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 703.941365] env[61868]: DEBUG nova.network.neutron [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.074923] env[61868]: DEBUG oslo_concurrency.lockutils [req-7e8cc533-2b7b-4c18-ad70-8a3f54472a26 req-c56c75dc-bb98-4b0a-a224-9d86a6af917c service nova] Releasing lock "refresh_cache-c0c9e61b-ab2b-47dc-997b-570204ab4c3b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.075387] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] Acquired lock "refresh_cache-c0c9e61b-ab2b-47dc-997b-570204ab4c3b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.075578] env[61868]: DEBUG nova.network.neutron [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 704.444466] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Releasing lock "refresh_cache-c979d105-12b7-485c-a136-55ffad02fe67" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.444743] env[61868]: DEBUG nova.compute.manager [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 704.444936] env[61868]: DEBUG nova.compute.manager [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 704.445110] env[61868]: DEBUG nova.network.neutron [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: c979d105-12b7-485c-a136-55ffad02fe67] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 704.459608] env[61868]: DEBUG nova.network.neutron [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 704.593032] env[61868]: DEBUG nova.network.neutron [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 704.650222] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bf59dea-0a05-4d86-af6a-c25bafdaca16 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.657950] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9835166a-b2e8-44b8-b2a1-c4ad302d928a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.663776] env[61868]: DEBUG nova.network.neutron [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.690289] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1f8cf4c-0b0b-4c98-b23e-13bed7042734 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.697976] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55e5e6fd-67d8-4689-8a4e-3b71a062f053 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.711747] env[61868]: DEBUG nova.compute.provider_tree [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 704.962081] env[61868]: DEBUG nova.network.neutron [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.993056] env[61868]: DEBUG nova.compute.manager [req-638f476c-faaf-46ad-8d35-84e32bceed3e req-05bb380f-d618-4342-af86-81395154c3c5 service nova] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Received event network-vif-deleted-48d4c453-40b0-4fb2-a83c-b145b3e20694 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 705.191018] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] Releasing lock "refresh_cache-c0c9e61b-ab2b-47dc-997b-570204ab4c3b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 705.191116] env[61868]: DEBUG nova.compute.manager [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 705.191256] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 705.191580] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4ac08de2-76f7-4843-862b-234cf40d124a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.201298] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39e28472-e580-474f-ad97-7bad9880ff53 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.214699] env[61868]: DEBUG nova.scheduler.client.report [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 705.222727] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c0c9e61b-ab2b-47dc-997b-570204ab4c3b could not be found. [ 705.222941] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 705.223130] env[61868]: INFO nova.compute.manager [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Took 0.03 seconds to destroy the instance on the hypervisor. [ 705.223872] env[61868]: DEBUG oslo.service.loopingcall [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 705.223872] env[61868]: DEBUG nova.compute.manager [-] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 705.223872] env[61868]: DEBUG nova.network.neutron [-] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 705.238324] env[61868]: DEBUG nova.network.neutron [-] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 705.465745] env[61868]: INFO nova.compute.manager [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] [instance: c979d105-12b7-485c-a136-55ffad02fe67] Took 1.02 seconds to deallocate network for instance. [ 705.719074] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.372s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.719640] env[61868]: DEBUG nova.compute.manager [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 705.722195] env[61868]: DEBUG oslo_concurrency.lockutils [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.085s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.740399] env[61868]: DEBUG nova.network.neutron [-] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.226331] env[61868]: DEBUG nova.compute.utils [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 706.230510] env[61868]: DEBUG nova.compute.manager [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 706.230757] env[61868]: DEBUG nova.network.neutron [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 706.242983] env[61868]: INFO nova.compute.manager [-] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Took 1.02 seconds to deallocate network for instance. [ 706.247844] env[61868]: DEBUG nova.compute.claims [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 706.248939] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.278686] env[61868]: DEBUG nova.policy [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '03ead7f08e934c4e9a1b39fbe08ed6e3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2cb049f497b14ac39a0316132f5802fc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 706.492766] env[61868]: INFO nova.scheduler.client.report [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Deleted allocations for instance c979d105-12b7-485c-a136-55ffad02fe67 [ 706.557562] env[61868]: DEBUG nova.network.neutron [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Successfully created port: 77c9c449-01ba-4bab-90f8-49033ffd0a67 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 706.571638] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5c57557-9b6f-474b-b253-dacd75277170 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.579293] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8c3b12b-2fc2-4143-92b6-d5f2d046cd5a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.608858] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c777aa6-8e59-4417-b720-600fb01dbeda {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.615931] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cf21eea-4435-48ad-b2cd-797bee2bfbbc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.628837] env[61868]: DEBUG nova.compute.provider_tree [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 706.733522] env[61868]: DEBUG nova.compute.manager [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 707.007521] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c314e496-459a-4772-a607-a9403c79e252 tempest-ServersAdminTestJSON-2044046595 tempest-ServersAdminTestJSON-2044046595-project-member] Lock "c979d105-12b7-485c-a136-55ffad02fe67" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 125.397s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.132148] env[61868]: DEBUG nova.scheduler.client.report [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 707.213966] env[61868]: DEBUG nova.compute.manager [req-7bbfa130-a5e3-4273-9b6b-4660f2317fd3 req-368fc2fa-9dc3-461b-8271-5935007dc0f2 service nova] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Received event network-changed-77c9c449-01ba-4bab-90f8-49033ffd0a67 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 707.214105] env[61868]: DEBUG nova.compute.manager [req-7bbfa130-a5e3-4273-9b6b-4660f2317fd3 req-368fc2fa-9dc3-461b-8271-5935007dc0f2 service nova] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Refreshing instance network info cache due to event network-changed-77c9c449-01ba-4bab-90f8-49033ffd0a67. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 707.214526] env[61868]: DEBUG oslo_concurrency.lockutils [req-7bbfa130-a5e3-4273-9b6b-4660f2317fd3 req-368fc2fa-9dc3-461b-8271-5935007dc0f2 service nova] Acquiring lock "refresh_cache-96694bb9-fbf2-4b71-9cb5-37f81b634992" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.214526] env[61868]: DEBUG oslo_concurrency.lockutils [req-7bbfa130-a5e3-4273-9b6b-4660f2317fd3 req-368fc2fa-9dc3-461b-8271-5935007dc0f2 service nova] Acquired lock "refresh_cache-96694bb9-fbf2-4b71-9cb5-37f81b634992" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.214687] env[61868]: DEBUG nova.network.neutron [req-7bbfa130-a5e3-4273-9b6b-4660f2317fd3 req-368fc2fa-9dc3-461b-8271-5935007dc0f2 service nova] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Refreshing network info cache for port 77c9c449-01ba-4bab-90f8-49033ffd0a67 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 707.432113] env[61868]: ERROR nova.compute.manager [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 77c9c449-01ba-4bab-90f8-49033ffd0a67, please check neutron logs for more information. [ 707.432113] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 707.432113] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 707.432113] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 707.432113] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 707.432113] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 707.432113] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 707.432113] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 707.432113] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 707.432113] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 707.432113] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 707.432113] env[61868]: ERROR nova.compute.manager raise self.value [ 707.432113] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 707.432113] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 707.432113] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 707.432113] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 707.432578] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 707.432578] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 707.432578] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 77c9c449-01ba-4bab-90f8-49033ffd0a67, please check neutron logs for more information. [ 707.432578] env[61868]: ERROR nova.compute.manager [ 707.432578] env[61868]: Traceback (most recent call last): [ 707.432578] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 707.432578] env[61868]: listener.cb(fileno) [ 707.432578] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 707.432578] env[61868]: result = function(*args, **kwargs) [ 707.432578] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 707.432578] env[61868]: return func(*args, **kwargs) [ 707.432578] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 707.432578] env[61868]: raise e [ 707.432578] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 707.432578] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 707.432578] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 707.432578] env[61868]: created_port_ids = self._update_ports_for_instance( [ 707.432578] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 707.432578] env[61868]: with excutils.save_and_reraise_exception(): [ 707.432578] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 707.432578] env[61868]: self.force_reraise() [ 707.432578] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 707.432578] env[61868]: raise self.value [ 707.432578] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 707.432578] env[61868]: updated_port = self._update_port( [ 707.432578] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 707.432578] env[61868]: _ensure_no_port_binding_failure(port) [ 707.432578] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 707.432578] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 707.433379] env[61868]: nova.exception.PortBindingFailed: Binding failed for port 77c9c449-01ba-4bab-90f8-49033ffd0a67, please check neutron logs for more information. [ 707.433379] env[61868]: Removing descriptor: 16 [ 707.513106] env[61868]: DEBUG nova.compute.manager [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 707.639020] env[61868]: DEBUG oslo_concurrency.lockutils [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.915s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.639020] env[61868]: ERROR nova.compute.manager [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ff91543d-3c5c-4222-8955-f78dcfa1da3a, please check neutron logs for more information. [ 707.639020] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Traceback (most recent call last): [ 707.639020] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 707.639020] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] self.driver.spawn(context, instance, image_meta, [ 707.639020] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 707.639020] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] self._vmops.spawn(context, instance, image_meta, injected_files, [ 707.639020] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 707.639020] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] vm_ref = self.build_virtual_machine(instance, [ 707.639430] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 707.639430] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] vif_infos = vmwarevif.get_vif_info(self._session, [ 707.639430] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 707.639430] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] for vif in network_info: [ 707.639430] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 707.639430] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] return self._sync_wrapper(fn, *args, **kwargs) [ 707.639430] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 707.639430] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] self.wait() [ 707.639430] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 707.639430] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] self[:] = self._gt.wait() [ 707.639430] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 707.639430] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] return self._exit_event.wait() [ 707.639430] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 707.639801] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] current.throw(*self._exc) [ 707.639801] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 707.639801] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] result = function(*args, **kwargs) [ 707.639801] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 707.639801] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] return func(*args, **kwargs) [ 707.639801] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 707.639801] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] raise e [ 707.639801] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 707.639801] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] nwinfo = self.network_api.allocate_for_instance( [ 707.639801] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 707.639801] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] created_port_ids = self._update_ports_for_instance( [ 707.639801] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 707.639801] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] with excutils.save_and_reraise_exception(): [ 707.640193] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 707.640193] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] self.force_reraise() [ 707.640193] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 707.640193] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] raise self.value [ 707.640193] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 707.640193] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] updated_port = self._update_port( [ 707.640193] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 707.640193] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] _ensure_no_port_binding_failure(port) [ 707.640193] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 707.640193] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] raise exception.PortBindingFailed(port_id=port['id']) [ 707.640193] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] nova.exception.PortBindingFailed: Binding failed for port ff91543d-3c5c-4222-8955-f78dcfa1da3a, please check neutron logs for more information. [ 707.640193] env[61868]: ERROR nova.compute.manager [instance: 0d80665a-c881-454a-90c7-96348e7fab74] [ 707.640538] env[61868]: DEBUG nova.compute.utils [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Binding failed for port ff91543d-3c5c-4222-8955-f78dcfa1da3a, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 707.644017] env[61868]: DEBUG nova.compute.manager [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Build of instance 0d80665a-c881-454a-90c7-96348e7fab74 was re-scheduled: Binding failed for port ff91543d-3c5c-4222-8955-f78dcfa1da3a, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 707.644017] env[61868]: DEBUG nova.compute.manager [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 707.644017] env[61868]: DEBUG oslo_concurrency.lockutils [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "refresh_cache-0d80665a-c881-454a-90c7-96348e7fab74" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.644017] env[61868]: DEBUG oslo_concurrency.lockutils [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquired lock "refresh_cache-0d80665a-c881-454a-90c7-96348e7fab74" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.644239] env[61868]: DEBUG nova.network.neutron [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 707.644239] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.090s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.645508] env[61868]: INFO nova.compute.claims [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 707.737980] env[61868]: DEBUG nova.network.neutron [req-7bbfa130-a5e3-4273-9b6b-4660f2317fd3 req-368fc2fa-9dc3-461b-8271-5935007dc0f2 service nova] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 707.744169] env[61868]: DEBUG nova.compute.manager [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 707.790693] env[61868]: DEBUG nova.virt.hardware [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 707.790992] env[61868]: DEBUG nova.virt.hardware [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 707.796234] env[61868]: DEBUG nova.virt.hardware [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 707.796553] env[61868]: DEBUG nova.virt.hardware [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 707.796783] env[61868]: DEBUG nova.virt.hardware [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 707.797029] env[61868]: DEBUG nova.virt.hardware [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 707.797361] env[61868]: DEBUG nova.virt.hardware [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 707.797619] env[61868]: DEBUG nova.virt.hardware [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 707.797875] env[61868]: DEBUG nova.virt.hardware [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 707.798122] env[61868]: DEBUG nova.virt.hardware [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 707.798345] env[61868]: DEBUG nova.virt.hardware [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 707.801364] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03dcd826-68b3-4262-86a5-1187c2d25df9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.817021] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4f64b98-84ea-433d-891d-e069d95a95bb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.829272] env[61868]: ERROR nova.compute.manager [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 77c9c449-01ba-4bab-90f8-49033ffd0a67, please check neutron logs for more information. [ 707.829272] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Traceback (most recent call last): [ 707.829272] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 707.829272] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] yield resources [ 707.829272] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 707.829272] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] self.driver.spawn(context, instance, image_meta, [ 707.829272] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 707.829272] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] self._vmops.spawn(context, instance, image_meta, injected_files, [ 707.829272] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 707.829272] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] vm_ref = self.build_virtual_machine(instance, [ 707.829272] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 707.829787] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] vif_infos = vmwarevif.get_vif_info(self._session, [ 707.829787] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 707.829787] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] for vif in network_info: [ 707.829787] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 707.829787] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] return self._sync_wrapper(fn, *args, **kwargs) [ 707.829787] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 707.829787] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] self.wait() [ 707.829787] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 707.829787] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] self[:] = self._gt.wait() [ 707.829787] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 707.829787] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] return self._exit_event.wait() [ 707.829787] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 707.829787] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] current.throw(*self._exc) [ 707.830230] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 707.830230] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] result = function(*args, **kwargs) [ 707.830230] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 707.830230] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] return func(*args, **kwargs) [ 707.830230] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 707.830230] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] raise e [ 707.830230] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 707.830230] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] nwinfo = self.network_api.allocate_for_instance( [ 707.830230] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 707.830230] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] created_port_ids = self._update_ports_for_instance( [ 707.830230] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 707.830230] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] with excutils.save_and_reraise_exception(): [ 707.830230] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 707.830730] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] self.force_reraise() [ 707.830730] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 707.830730] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] raise self.value [ 707.830730] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 707.830730] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] updated_port = self._update_port( [ 707.830730] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 707.830730] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] _ensure_no_port_binding_failure(port) [ 707.830730] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 707.830730] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] raise exception.PortBindingFailed(port_id=port['id']) [ 707.830730] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] nova.exception.PortBindingFailed: Binding failed for port 77c9c449-01ba-4bab-90f8-49033ffd0a67, please check neutron logs for more information. [ 707.830730] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] [ 707.832987] env[61868]: INFO nova.compute.manager [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Terminating instance [ 707.833917] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquiring lock "refresh_cache-96694bb9-fbf2-4b71-9cb5-37f81b634992" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.859335] env[61868]: DEBUG nova.network.neutron [req-7bbfa130-a5e3-4273-9b6b-4660f2317fd3 req-368fc2fa-9dc3-461b-8271-5935007dc0f2 service nova] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.047365] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 708.171773] env[61868]: DEBUG nova.network.neutron [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 708.245602] env[61868]: DEBUG nova.network.neutron [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.362269] env[61868]: DEBUG oslo_concurrency.lockutils [req-7bbfa130-a5e3-4273-9b6b-4660f2317fd3 req-368fc2fa-9dc3-461b-8271-5935007dc0f2 service nova] Releasing lock "refresh_cache-96694bb9-fbf2-4b71-9cb5-37f81b634992" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.364966] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquired lock "refresh_cache-96694bb9-fbf2-4b71-9cb5-37f81b634992" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.364966] env[61868]: DEBUG nova.network.neutron [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 708.749287] env[61868]: DEBUG oslo_concurrency.lockutils [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Releasing lock "refresh_cache-0d80665a-c881-454a-90c7-96348e7fab74" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.749287] env[61868]: DEBUG nova.compute.manager [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 708.749287] env[61868]: DEBUG nova.compute.manager [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 708.749287] env[61868]: DEBUG nova.network.neutron [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 708.763603] env[61868]: DEBUG nova.network.neutron [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 708.882469] env[61868]: DEBUG nova.network.neutron [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 708.987248] env[61868]: DEBUG nova.network.neutron [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.014644] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c10f7b48-928d-448e-a55c-1fb936a891ce {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.025196] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-184e4448-bfec-4c36-9d69-f65913a56336 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.058508] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d14f80c0-040f-4ac7-8d20-2e190062a49a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.066279] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d46f7505-2b09-45d6-8c29-a280752f75ea {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.083913] env[61868]: DEBUG nova.compute.provider_tree [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 709.242850] env[61868]: DEBUG nova.compute.manager [req-f87cc273-6c0b-4947-ae9d-4f07844437bd req-27145235-2d94-44d6-9cf1-c13fecba1fbb service nova] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Received event network-vif-deleted-77c9c449-01ba-4bab-90f8-49033ffd0a67 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 709.266730] env[61868]: DEBUG nova.network.neutron [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.487334] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Releasing lock "refresh_cache-96694bb9-fbf2-4b71-9cb5-37f81b634992" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 709.487759] env[61868]: DEBUG nova.compute.manager [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 709.487956] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 709.488277] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c7799b0f-35ee-43bc-b3ca-67a2809621c4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.496897] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f26ec7fb-7a31-4511-9114-6812591c37d8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.519343] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 96694bb9-fbf2-4b71-9cb5-37f81b634992 could not be found. [ 709.519546] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 709.519725] env[61868]: INFO nova.compute.manager [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Took 0.03 seconds to destroy the instance on the hypervisor. [ 709.519987] env[61868]: DEBUG oslo.service.loopingcall [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 709.520214] env[61868]: DEBUG nova.compute.manager [-] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 709.520304] env[61868]: DEBUG nova.network.neutron [-] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 709.535146] env[61868]: DEBUG nova.network.neutron [-] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 709.586495] env[61868]: DEBUG nova.scheduler.client.report [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 709.769416] env[61868]: INFO nova.compute.manager [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 0d80665a-c881-454a-90c7-96348e7fab74] Took 1.02 seconds to deallocate network for instance. [ 710.037342] env[61868]: DEBUG nova.network.neutron [-] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.091737] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.448s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.093276] env[61868]: DEBUG nova.compute.manager [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 710.095719] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.929s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.540277] env[61868]: INFO nova.compute.manager [-] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Took 1.02 seconds to deallocate network for instance. [ 710.542789] env[61868]: DEBUG nova.compute.claims [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 710.542976] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.605168] env[61868]: DEBUG nova.compute.utils [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 710.607485] env[61868]: DEBUG nova.compute.manager [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 710.607485] env[61868]: DEBUG nova.network.neutron [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 710.671263] env[61868]: DEBUG nova.policy [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '379bc4b2a0e34a779eb064dd18f02a08', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7471600a510e4feeb0c819d0cca5eb48', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 710.810467] env[61868]: INFO nova.scheduler.client.report [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Deleted allocations for instance 0d80665a-c881-454a-90c7-96348e7fab74 [ 711.008950] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69220a0c-030d-4fdb-bc70-e607e9da52a7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.017525] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e07cc75b-0cc7-47a8-9b12-d83659d7378f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.051050] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d98de686-05aa-4761-87cb-56aa2ee0b8fe {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.058646] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-728ed669-c106-4c8e-8c99-aa51b94b88df {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.075216] env[61868]: DEBUG nova.compute.provider_tree [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 711.086216] env[61868]: DEBUG nova.network.neutron [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Successfully created port: 80ca8e4e-c979-4622-898c-7901cb81d1dc {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 711.109509] env[61868]: DEBUG nova.compute.manager [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 711.327647] env[61868]: DEBUG oslo_concurrency.lockutils [None req-152a9053-d0ff-4671-a68d-e6076c926995 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "0d80665a-c881-454a-90c7-96348e7fab74" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 129.009s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.440605] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Acquiring lock "2c00210e-3b6a-4041-9e8e-7f40ef23d716" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 711.440605] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Lock "2c00210e-3b6a-4041-9e8e-7f40ef23d716" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 711.577808] env[61868]: DEBUG nova.scheduler.client.report [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 711.829086] env[61868]: DEBUG nova.compute.manager [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 711.966035] env[61868]: DEBUG nova.compute.manager [req-153fde78-b56f-4f18-944e-ed152f6b4350 req-b2b25968-e5c6-47d8-8338-0fa933df7876 service nova] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Received event network-changed-80ca8e4e-c979-4622-898c-7901cb81d1dc {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 711.966265] env[61868]: DEBUG nova.compute.manager [req-153fde78-b56f-4f18-944e-ed152f6b4350 req-b2b25968-e5c6-47d8-8338-0fa933df7876 service nova] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Refreshing instance network info cache due to event network-changed-80ca8e4e-c979-4622-898c-7901cb81d1dc. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 711.966634] env[61868]: DEBUG oslo_concurrency.lockutils [req-153fde78-b56f-4f18-944e-ed152f6b4350 req-b2b25968-e5c6-47d8-8338-0fa933df7876 service nova] Acquiring lock "refresh_cache-cf7b7da2-380e-4a97-b142-0dd9e5ad3c79" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.966802] env[61868]: DEBUG oslo_concurrency.lockutils [req-153fde78-b56f-4f18-944e-ed152f6b4350 req-b2b25968-e5c6-47d8-8338-0fa933df7876 service nova] Acquired lock "refresh_cache-cf7b7da2-380e-4a97-b142-0dd9e5ad3c79" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.967053] env[61868]: DEBUG nova.network.neutron [req-153fde78-b56f-4f18-944e-ed152f6b4350 req-b2b25968-e5c6-47d8-8338-0fa933df7876 service nova] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Refreshing network info cache for port 80ca8e4e-c979-4622-898c-7901cb81d1dc {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 712.042634] env[61868]: ERROR nova.compute.manager [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 80ca8e4e-c979-4622-898c-7901cb81d1dc, please check neutron logs for more information. [ 712.042634] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 712.042634] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.042634] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 712.042634] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 712.042634] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 712.042634] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 712.042634] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 712.042634] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.042634] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 712.042634] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.042634] env[61868]: ERROR nova.compute.manager raise self.value [ 712.042634] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 712.042634] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 712.042634] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.042634] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 712.043228] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.043228] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 712.043228] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 80ca8e4e-c979-4622-898c-7901cb81d1dc, please check neutron logs for more information. [ 712.043228] env[61868]: ERROR nova.compute.manager [ 712.043228] env[61868]: Traceback (most recent call last): [ 712.043228] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 712.043228] env[61868]: listener.cb(fileno) [ 712.043228] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 712.043228] env[61868]: result = function(*args, **kwargs) [ 712.043228] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 712.043228] env[61868]: return func(*args, **kwargs) [ 712.043228] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 712.043228] env[61868]: raise e [ 712.043228] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.043228] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 712.043228] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 712.043228] env[61868]: created_port_ids = self._update_ports_for_instance( [ 712.043228] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 712.043228] env[61868]: with excutils.save_and_reraise_exception(): [ 712.043228] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.043228] env[61868]: self.force_reraise() [ 712.043228] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.043228] env[61868]: raise self.value [ 712.043228] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 712.043228] env[61868]: updated_port = self._update_port( [ 712.043228] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.043228] env[61868]: _ensure_no_port_binding_failure(port) [ 712.043228] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.043228] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 712.044091] env[61868]: nova.exception.PortBindingFailed: Binding failed for port 80ca8e4e-c979-4622-898c-7901cb81d1dc, please check neutron logs for more information. [ 712.044091] env[61868]: Removing descriptor: 16 [ 712.082881] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.987s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.083577] env[61868]: ERROR nova.compute.manager [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6a100ab0-5a39-4a17-9ed1-87a211f37c0b, please check neutron logs for more information. [ 712.083577] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Traceback (most recent call last): [ 712.083577] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 712.083577] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] self.driver.spawn(context, instance, image_meta, [ 712.083577] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 712.083577] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 712.083577] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 712.083577] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] vm_ref = self.build_virtual_machine(instance, [ 712.083577] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 712.083577] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] vif_infos = vmwarevif.get_vif_info(self._session, [ 712.083577] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 712.083919] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] for vif in network_info: [ 712.083919] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 712.083919] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] return self._sync_wrapper(fn, *args, **kwargs) [ 712.083919] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 712.083919] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] self.wait() [ 712.083919] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 712.083919] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] self[:] = self._gt.wait() [ 712.083919] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 712.083919] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] return self._exit_event.wait() [ 712.083919] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 712.083919] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] current.throw(*self._exc) [ 712.083919] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 712.083919] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] result = function(*args, **kwargs) [ 712.084945] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 712.084945] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] return func(*args, **kwargs) [ 712.084945] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 712.084945] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] raise e [ 712.084945] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.084945] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] nwinfo = self.network_api.allocate_for_instance( [ 712.084945] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 712.084945] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] created_port_ids = self._update_ports_for_instance( [ 712.084945] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 712.084945] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] with excutils.save_and_reraise_exception(): [ 712.084945] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.084945] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] self.force_reraise() [ 712.084945] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.085303] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] raise self.value [ 712.085303] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 712.085303] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] updated_port = self._update_port( [ 712.085303] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.085303] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] _ensure_no_port_binding_failure(port) [ 712.085303] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.085303] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] raise exception.PortBindingFailed(port_id=port['id']) [ 712.085303] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] nova.exception.PortBindingFailed: Binding failed for port 6a100ab0-5a39-4a17-9ed1-87a211f37c0b, please check neutron logs for more information. [ 712.085303] env[61868]: ERROR nova.compute.manager [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] [ 712.085303] env[61868]: DEBUG nova.compute.utils [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Binding failed for port 6a100ab0-5a39-4a17-9ed1-87a211f37c0b, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 712.092036] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.235s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.092036] env[61868]: INFO nova.compute.claims [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 712.094817] env[61868]: DEBUG nova.compute.manager [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Build of instance b6118327-0376-4b05-aee3-cb56d9cdf0b2 was re-scheduled: Binding failed for port 6a100ab0-5a39-4a17-9ed1-87a211f37c0b, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 712.095554] env[61868]: DEBUG nova.compute.manager [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 712.095889] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquiring lock "refresh_cache-b6118327-0376-4b05-aee3-cb56d9cdf0b2" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.096170] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquired lock "refresh_cache-b6118327-0376-4b05-aee3-cb56d9cdf0b2" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.096437] env[61868]: DEBUG nova.network.neutron [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 712.118891] env[61868]: DEBUG nova.compute.manager [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 712.150296] env[61868]: DEBUG nova.virt.hardware [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 712.150558] env[61868]: DEBUG nova.virt.hardware [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 712.150782] env[61868]: DEBUG nova.virt.hardware [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 712.150977] env[61868]: DEBUG nova.virt.hardware [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 712.151139] env[61868]: DEBUG nova.virt.hardware [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 712.151313] env[61868]: DEBUG nova.virt.hardware [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 712.151530] env[61868]: DEBUG nova.virt.hardware [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 712.151649] env[61868]: DEBUG nova.virt.hardware [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 712.151856] env[61868]: DEBUG nova.virt.hardware [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 712.152034] env[61868]: DEBUG nova.virt.hardware [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 712.152213] env[61868]: DEBUG nova.virt.hardware [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 712.153294] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-347b9ef8-9d78-41df-83d8-62e6a370d3c7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.163029] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-673a3a96-b0ba-45e2-ac2d-cdde16c136dd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.177695] env[61868]: ERROR nova.compute.manager [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 80ca8e4e-c979-4622-898c-7901cb81d1dc, please check neutron logs for more information. [ 712.177695] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Traceback (most recent call last): [ 712.177695] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 712.177695] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] yield resources [ 712.177695] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 712.177695] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] self.driver.spawn(context, instance, image_meta, [ 712.177695] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 712.177695] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] self._vmops.spawn(context, instance, image_meta, injected_files, [ 712.177695] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 712.177695] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] vm_ref = self.build_virtual_machine(instance, [ 712.177695] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 712.178139] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] vif_infos = vmwarevif.get_vif_info(self._session, [ 712.178139] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 712.178139] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] for vif in network_info: [ 712.178139] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 712.178139] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] return self._sync_wrapper(fn, *args, **kwargs) [ 712.178139] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 712.178139] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] self.wait() [ 712.178139] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 712.178139] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] self[:] = self._gt.wait() [ 712.178139] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 712.178139] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] return self._exit_event.wait() [ 712.178139] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 712.178139] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] current.throw(*self._exc) [ 712.178522] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 712.178522] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] result = function(*args, **kwargs) [ 712.178522] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 712.178522] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] return func(*args, **kwargs) [ 712.178522] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 712.178522] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] raise e [ 712.178522] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.178522] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] nwinfo = self.network_api.allocate_for_instance( [ 712.178522] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 712.178522] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] created_port_ids = self._update_ports_for_instance( [ 712.178522] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 712.178522] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] with excutils.save_and_reraise_exception(): [ 712.178522] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.178920] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] self.force_reraise() [ 712.178920] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.178920] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] raise self.value [ 712.178920] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 712.178920] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] updated_port = self._update_port( [ 712.178920] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.178920] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] _ensure_no_port_binding_failure(port) [ 712.178920] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.178920] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] raise exception.PortBindingFailed(port_id=port['id']) [ 712.178920] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] nova.exception.PortBindingFailed: Binding failed for port 80ca8e4e-c979-4622-898c-7901cb81d1dc, please check neutron logs for more information. [ 712.178920] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] [ 712.178920] env[61868]: INFO nova.compute.manager [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Terminating instance [ 712.180295] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Acquiring lock "refresh_cache-cf7b7da2-380e-4a97-b142-0dd9e5ad3c79" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.348285] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.486891] env[61868]: DEBUG nova.network.neutron [req-153fde78-b56f-4f18-944e-ed152f6b4350 req-b2b25968-e5c6-47d8-8338-0fa933df7876 service nova] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.597422] env[61868]: DEBUG nova.network.neutron [req-153fde78-b56f-4f18-944e-ed152f6b4350 req-b2b25968-e5c6-47d8-8338-0fa933df7876 service nova] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.618224] env[61868]: DEBUG nova.network.neutron [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.699393] env[61868]: DEBUG nova.network.neutron [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.101621] env[61868]: DEBUG oslo_concurrency.lockutils [req-153fde78-b56f-4f18-944e-ed152f6b4350 req-b2b25968-e5c6-47d8-8338-0fa933df7876 service nova] Releasing lock "refresh_cache-cf7b7da2-380e-4a97-b142-0dd9e5ad3c79" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.102039] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Acquired lock "refresh_cache-cf7b7da2-380e-4a97-b142-0dd9e5ad3c79" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.102227] env[61868]: DEBUG nova.network.neutron [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 713.110263] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Acquiring lock "d35953d8-692d-498e-baf5-96ef381ce12b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 713.110892] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Lock "d35953d8-692d-498e-baf5-96ef381ce12b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.202671] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Releasing lock "refresh_cache-b6118327-0376-4b05-aee3-cb56d9cdf0b2" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.202671] env[61868]: DEBUG nova.compute.manager [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 713.202671] env[61868]: DEBUG nova.compute.manager [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 713.202847] env[61868]: DEBUG nova.network.neutron [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 713.225548] env[61868]: DEBUG nova.network.neutron [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 713.488685] env[61868]: DEBUG oslo_concurrency.lockutils [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Acquiring lock "02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 713.488915] env[61868]: DEBUG oslo_concurrency.lockutils [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Lock "02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.491725] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "a16a726b-9673-4de7-9188-19c608a7dc0b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 713.491922] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "a16a726b-9673-4de7-9188-19c608a7dc0b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.513118] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41599ffa-a5be-4d61-9066-e56be222cfb8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.519592] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab3b88d5-b9fc-421e-9691-4a4bf54161cb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.551357] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ec03fa7-8ee0-47b5-984c-a8175486130e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.558415] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9c90130-ae2e-4c5b-9a4a-e8c128bc7c83 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.571413] env[61868]: DEBUG nova.compute.provider_tree [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 713.623484] env[61868]: DEBUG nova.network.neutron [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 713.703282] env[61868]: DEBUG nova.network.neutron [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.728202] env[61868]: DEBUG nova.network.neutron [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.074167] env[61868]: DEBUG nova.scheduler.client.report [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 714.110426] env[61868]: DEBUG nova.compute.manager [req-0f16a06e-0335-412e-8bec-c87a33d8dcd5 req-202115dc-46da-4ff0-83c4-4cd115e73679 service nova] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Received event network-vif-deleted-80ca8e4e-c979-4622-898c-7901cb81d1dc {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 714.206507] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Releasing lock "refresh_cache-cf7b7da2-380e-4a97-b142-0dd9e5ad3c79" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.207089] env[61868]: DEBUG nova.compute.manager [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 714.207257] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 714.207617] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-64b12ff2-9634-4e29-8a12-49833bc360ba {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.217641] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-665ae4a7-8b28-4d24-9b01-e7897c0ed89e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.231638] env[61868]: INFO nova.compute.manager [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: b6118327-0376-4b05-aee3-cb56d9cdf0b2] Took 1.03 seconds to deallocate network for instance. [ 714.241431] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance cf7b7da2-380e-4a97-b142-0dd9e5ad3c79 could not be found. [ 714.241736] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 714.242010] env[61868]: INFO nova.compute.manager [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Took 0.03 seconds to destroy the instance on the hypervisor. [ 714.242305] env[61868]: DEBUG oslo.service.loopingcall [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 714.242535] env[61868]: DEBUG nova.compute.manager [-] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 714.242656] env[61868]: DEBUG nova.network.neutron [-] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 714.259745] env[61868]: DEBUG nova.network.neutron [-] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 714.579698] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.491s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.580287] env[61868]: DEBUG nova.compute.manager [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 714.582954] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.581s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 714.584350] env[61868]: INFO nova.compute.claims [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 714.762793] env[61868]: DEBUG nova.network.neutron [-] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.089183] env[61868]: DEBUG nova.compute.utils [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 715.092279] env[61868]: DEBUG nova.compute.manager [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 715.094331] env[61868]: DEBUG nova.network.neutron [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] [instance: 460255fb-6589-481b-9770-0a2b548c6194] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 715.134682] env[61868]: DEBUG nova.policy [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3c55be09308244c681260c15bc3b348d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd285246f7ed9464f8bfab90e2427fa0f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 715.264888] env[61868]: INFO nova.scheduler.client.report [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Deleted allocations for instance b6118327-0376-4b05-aee3-cb56d9cdf0b2 [ 715.271053] env[61868]: INFO nova.compute.manager [-] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Took 1.03 seconds to deallocate network for instance. [ 715.274310] env[61868]: DEBUG nova.compute.claims [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 715.274310] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 715.430478] env[61868]: DEBUG nova.network.neutron [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Successfully created port: 6f2211c4-2123-49a9-b287-3b10300c560d {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 715.595471] env[61868]: DEBUG nova.compute.manager [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 715.774268] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Lock "b6118327-0376-4b05-aee3-cb56d9cdf0b2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 130.366s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.930999] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4400e436-25ef-4ed4-9ac4-a91b8b10d7b1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.938581] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78809707-f9c1-46e8-afdf-0a8c6adc98f9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.968184] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4358333-9fe5-4e2a-928d-0059d8a7c34f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.975362] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8aa4f5e-d52c-436e-9b62-5561ac70595f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.989399] env[61868]: DEBUG nova.compute.provider_tree [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Updating inventory in ProviderTree for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 716.225511] env[61868]: DEBUG nova.compute.manager [req-218425c2-cae5-49c9-8c6e-6570aaf0ab02 req-acd82cdb-cace-4d0b-a7b0-33909be2f5eb service nova] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Received event network-changed-6f2211c4-2123-49a9-b287-3b10300c560d {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 716.229110] env[61868]: DEBUG nova.compute.manager [req-218425c2-cae5-49c9-8c6e-6570aaf0ab02 req-acd82cdb-cace-4d0b-a7b0-33909be2f5eb service nova] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Refreshing instance network info cache due to event network-changed-6f2211c4-2123-49a9-b287-3b10300c560d. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 716.229110] env[61868]: DEBUG oslo_concurrency.lockutils [req-218425c2-cae5-49c9-8c6e-6570aaf0ab02 req-acd82cdb-cace-4d0b-a7b0-33909be2f5eb service nova] Acquiring lock "refresh_cache-460255fb-6589-481b-9770-0a2b548c6194" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.229110] env[61868]: DEBUG oslo_concurrency.lockutils [req-218425c2-cae5-49c9-8c6e-6570aaf0ab02 req-acd82cdb-cace-4d0b-a7b0-33909be2f5eb service nova] Acquired lock "refresh_cache-460255fb-6589-481b-9770-0a2b548c6194" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.229110] env[61868]: DEBUG nova.network.neutron [req-218425c2-cae5-49c9-8c6e-6570aaf0ab02 req-acd82cdb-cace-4d0b-a7b0-33909be2f5eb service nova] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Refreshing network info cache for port 6f2211c4-2123-49a9-b287-3b10300c560d {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 716.277643] env[61868]: DEBUG nova.compute.manager [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 716.411216] env[61868]: ERROR nova.compute.manager [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6f2211c4-2123-49a9-b287-3b10300c560d, please check neutron logs for more information. [ 716.411216] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 716.411216] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 716.411216] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 716.411216] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 716.411216] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 716.411216] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 716.411216] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 716.411216] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 716.411216] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 716.411216] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 716.411216] env[61868]: ERROR nova.compute.manager raise self.value [ 716.411216] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 716.411216] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 716.411216] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 716.411216] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 716.411756] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 716.411756] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 716.411756] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6f2211c4-2123-49a9-b287-3b10300c560d, please check neutron logs for more information. [ 716.411756] env[61868]: ERROR nova.compute.manager [ 716.411756] env[61868]: Traceback (most recent call last): [ 716.411756] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 716.411756] env[61868]: listener.cb(fileno) [ 716.411756] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 716.411756] env[61868]: result = function(*args, **kwargs) [ 716.411756] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 716.411756] env[61868]: return func(*args, **kwargs) [ 716.411756] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 716.411756] env[61868]: raise e [ 716.411756] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 716.411756] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 716.411756] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 716.411756] env[61868]: created_port_ids = self._update_ports_for_instance( [ 716.411756] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 716.411756] env[61868]: with excutils.save_and_reraise_exception(): [ 716.411756] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 716.411756] env[61868]: self.force_reraise() [ 716.411756] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 716.411756] env[61868]: raise self.value [ 716.411756] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 716.411756] env[61868]: updated_port = self._update_port( [ 716.411756] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 716.411756] env[61868]: _ensure_no_port_binding_failure(port) [ 716.411756] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 716.411756] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 716.412605] env[61868]: nova.exception.PortBindingFailed: Binding failed for port 6f2211c4-2123-49a9-b287-3b10300c560d, please check neutron logs for more information. [ 716.412605] env[61868]: Removing descriptor: 16 [ 716.511183] env[61868]: ERROR nova.scheduler.client.report [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [req-6d949fe5-ee69-43c6-abf4-f8c0c5f31672] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 83e67721-2ac9-4a23-aa31-82aca86979c8. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-6d949fe5-ee69-43c6-abf4-f8c0c5f31672"}]} [ 716.528871] env[61868]: DEBUG nova.scheduler.client.report [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Refreshing inventories for resource provider 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 716.545652] env[61868]: DEBUG nova.scheduler.client.report [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Updating ProviderTree inventory for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 716.545917] env[61868]: DEBUG nova.compute.provider_tree [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Updating inventory in ProviderTree for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 716.561879] env[61868]: DEBUG nova.scheduler.client.report [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Refreshing aggregate associations for resource provider 83e67721-2ac9-4a23-aa31-82aca86979c8, aggregates: None {{(pid=61868) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 716.581343] env[61868]: DEBUG nova.scheduler.client.report [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Refreshing trait associations for resource provider 83e67721-2ac9-4a23-aa31-82aca86979c8, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61868) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 716.606881] env[61868]: DEBUG nova.compute.manager [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 716.633395] env[61868]: DEBUG nova.virt.hardware [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 716.634424] env[61868]: DEBUG nova.virt.hardware [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 716.634424] env[61868]: DEBUG nova.virt.hardware [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 716.634424] env[61868]: DEBUG nova.virt.hardware [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 716.634424] env[61868]: DEBUG nova.virt.hardware [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 716.634424] env[61868]: DEBUG nova.virt.hardware [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 716.634623] env[61868]: DEBUG nova.virt.hardware [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 716.634956] env[61868]: DEBUG nova.virt.hardware [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 716.635328] env[61868]: DEBUG nova.virt.hardware [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 716.635388] env[61868]: DEBUG nova.virt.hardware [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 716.635589] env[61868]: DEBUG nova.virt.hardware [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 716.636667] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46243ee6-f8b2-4376-9624-80e004dba98a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.647268] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78d7c31e-e297-4b26-9176-fdcae06ecd5b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.663140] env[61868]: ERROR nova.compute.manager [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6f2211c4-2123-49a9-b287-3b10300c560d, please check neutron logs for more information. [ 716.663140] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] Traceback (most recent call last): [ 716.663140] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 716.663140] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] yield resources [ 716.663140] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 716.663140] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] self.driver.spawn(context, instance, image_meta, [ 716.663140] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 716.663140] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] self._vmops.spawn(context, instance, image_meta, injected_files, [ 716.663140] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 716.663140] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] vm_ref = self.build_virtual_machine(instance, [ 716.663140] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 716.663504] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] vif_infos = vmwarevif.get_vif_info(self._session, [ 716.663504] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 716.663504] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] for vif in network_info: [ 716.663504] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 716.663504] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] return self._sync_wrapper(fn, *args, **kwargs) [ 716.663504] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 716.663504] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] self.wait() [ 716.663504] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 716.663504] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] self[:] = self._gt.wait() [ 716.663504] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 716.663504] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] return self._exit_event.wait() [ 716.663504] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 716.663504] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] current.throw(*self._exc) [ 716.663887] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 716.663887] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] result = function(*args, **kwargs) [ 716.663887] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 716.663887] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] return func(*args, **kwargs) [ 716.663887] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 716.663887] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] raise e [ 716.663887] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 716.663887] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] nwinfo = self.network_api.allocate_for_instance( [ 716.663887] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 716.663887] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] created_port_ids = self._update_ports_for_instance( [ 716.663887] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 716.663887] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] with excutils.save_and_reraise_exception(): [ 716.663887] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 716.664335] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] self.force_reraise() [ 716.664335] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 716.664335] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] raise self.value [ 716.664335] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 716.664335] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] updated_port = self._update_port( [ 716.664335] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 716.664335] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] _ensure_no_port_binding_failure(port) [ 716.664335] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 716.664335] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] raise exception.PortBindingFailed(port_id=port['id']) [ 716.664335] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] nova.exception.PortBindingFailed: Binding failed for port 6f2211c4-2123-49a9-b287-3b10300c560d, please check neutron logs for more information. [ 716.664335] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] [ 716.664335] env[61868]: INFO nova.compute.manager [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Terminating instance [ 716.667091] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] Acquiring lock "refresh_cache-460255fb-6589-481b-9770-0a2b548c6194" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.742970] env[61868]: DEBUG nova.network.neutron [req-218425c2-cae5-49c9-8c6e-6570aaf0ab02 req-acd82cdb-cace-4d0b-a7b0-33909be2f5eb service nova] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 716.794682] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.818377] env[61868]: DEBUG nova.network.neutron [req-218425c2-cae5-49c9-8c6e-6570aaf0ab02 req-acd82cdb-cace-4d0b-a7b0-33909be2f5eb service nova] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.887988] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8beb37a-7a72-4081-b504-8280314aa797 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.895589] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6afe424-8681-455b-88a3-1a39bcdeb7d2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.927081] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6e3aff7-a1bc-43b6-ac27-3891ed354207 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.934315] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dce8b22-357d-4515-8458-bd6303782455 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.947924] env[61868]: DEBUG nova.compute.provider_tree [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Updating inventory in ProviderTree for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 717.321270] env[61868]: DEBUG oslo_concurrency.lockutils [req-218425c2-cae5-49c9-8c6e-6570aaf0ab02 req-acd82cdb-cace-4d0b-a7b0-33909be2f5eb service nova] Releasing lock "refresh_cache-460255fb-6589-481b-9770-0a2b548c6194" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.321662] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] Acquired lock "refresh_cache-460255fb-6589-481b-9770-0a2b548c6194" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.321846] env[61868]: DEBUG nova.network.neutron [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 717.481542] env[61868]: DEBUG nova.scheduler.client.report [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Updated inventory for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 with generation 65 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 717.481834] env[61868]: DEBUG nova.compute.provider_tree [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Updating resource provider 83e67721-2ac9-4a23-aa31-82aca86979c8 generation from 65 to 66 during operation: update_inventory {{(pid=61868) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 717.482038] env[61868]: DEBUG nova.compute.provider_tree [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Updating inventory in ProviderTree for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 717.840009] env[61868]: DEBUG nova.network.neutron [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 717.917601] env[61868]: DEBUG nova.network.neutron [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.986743] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.404s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.987258] env[61868]: DEBUG nova.compute.manager [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 717.990011] env[61868]: DEBUG oslo_concurrency.lockutils [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.165s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.991399] env[61868]: INFO nova.compute.claims [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 718.252757] env[61868]: DEBUG nova.compute.manager [req-9efc29d0-9d9f-4720-8518-a7f24b137636 req-071a1fd6-cc20-4447-bab2-af838478d4f3 service nova] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Received event network-vif-deleted-6f2211c4-2123-49a9-b287-3b10300c560d {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 718.419959] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] Releasing lock "refresh_cache-460255fb-6589-481b-9770-0a2b548c6194" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.420454] env[61868]: DEBUG nova.compute.manager [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 718.420671] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 718.421018] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-83a7ca39-7448-4ec6-b253-6f6b6d095950 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.431028] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2755a76-55a1-4d2b-82c2-281fc1f23934 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.457884] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 460255fb-6589-481b-9770-0a2b548c6194 could not be found. [ 718.458129] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 718.458309] env[61868]: INFO nova.compute.manager [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Took 0.04 seconds to destroy the instance on the hypervisor. [ 718.458628] env[61868]: DEBUG oslo.service.loopingcall [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 718.458868] env[61868]: DEBUG nova.compute.manager [-] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 718.458957] env[61868]: DEBUG nova.network.neutron [-] [instance: 460255fb-6589-481b-9770-0a2b548c6194] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 718.474167] env[61868]: DEBUG nova.network.neutron [-] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 718.498091] env[61868]: DEBUG nova.compute.utils [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 718.500917] env[61868]: DEBUG nova.compute.manager [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 718.500917] env[61868]: DEBUG nova.network.neutron [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 718.537691] env[61868]: DEBUG nova.policy [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e02ffdc140274733a081fd53c4acc202', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a60c3c20950e4156b64c3b4c61b9f0f2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 718.795268] env[61868]: DEBUG nova.network.neutron [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Successfully created port: 9fe37d4c-b367-48ea-b471-3a851ec07246 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 718.977037] env[61868]: DEBUG nova.network.neutron [-] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.003453] env[61868]: DEBUG nova.compute.manager [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 719.327378] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7f4f6d6-5489-459b-8079-15ec709a52f1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.338261] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4e29bf7-627f-4d3f-9f96-c4749f815251 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.388170] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed1a4a17-185e-4c1c-8908-a1112bdf53d2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.397787] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc2d3a2f-2464-4d35-8adf-fde14f42a240 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.412691] env[61868]: DEBUG nova.compute.provider_tree [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 719.479543] env[61868]: INFO nova.compute.manager [-] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Took 1.02 seconds to deallocate network for instance. [ 719.483993] env[61868]: DEBUG nova.compute.claims [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 719.484153] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.631514] env[61868]: ERROR nova.compute.manager [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9fe37d4c-b367-48ea-b471-3a851ec07246, please check neutron logs for more information. [ 719.631514] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 719.631514] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 719.631514] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 719.631514] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 719.631514] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 719.631514] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 719.631514] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 719.631514] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 719.631514] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 719.631514] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 719.631514] env[61868]: ERROR nova.compute.manager raise self.value [ 719.631514] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 719.631514] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 719.631514] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 719.631514] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 719.632269] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 719.632269] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 719.632269] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9fe37d4c-b367-48ea-b471-3a851ec07246, please check neutron logs for more information. [ 719.632269] env[61868]: ERROR nova.compute.manager [ 719.632269] env[61868]: Traceback (most recent call last): [ 719.632269] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 719.632269] env[61868]: listener.cb(fileno) [ 719.632269] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 719.632269] env[61868]: result = function(*args, **kwargs) [ 719.632269] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 719.632269] env[61868]: return func(*args, **kwargs) [ 719.632269] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 719.632269] env[61868]: raise e [ 719.632269] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 719.632269] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 719.632269] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 719.632269] env[61868]: created_port_ids = self._update_ports_for_instance( [ 719.632269] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 719.632269] env[61868]: with excutils.save_and_reraise_exception(): [ 719.632269] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 719.632269] env[61868]: self.force_reraise() [ 719.632269] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 719.632269] env[61868]: raise self.value [ 719.632269] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 719.632269] env[61868]: updated_port = self._update_port( [ 719.632269] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 719.632269] env[61868]: _ensure_no_port_binding_failure(port) [ 719.632269] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 719.632269] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 719.633358] env[61868]: nova.exception.PortBindingFailed: Binding failed for port 9fe37d4c-b367-48ea-b471-3a851ec07246, please check neutron logs for more information. [ 719.633358] env[61868]: Removing descriptor: 16 [ 719.916342] env[61868]: DEBUG nova.scheduler.client.report [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 720.013626] env[61868]: DEBUG nova.compute.manager [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 720.038319] env[61868]: DEBUG nova.virt.hardware [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 720.038574] env[61868]: DEBUG nova.virt.hardware [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 720.038730] env[61868]: DEBUG nova.virt.hardware [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 720.038975] env[61868]: DEBUG nova.virt.hardware [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 720.039070] env[61868]: DEBUG nova.virt.hardware [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 720.039224] env[61868]: DEBUG nova.virt.hardware [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 720.039430] env[61868]: DEBUG nova.virt.hardware [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 720.039589] env[61868]: DEBUG nova.virt.hardware [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 720.039754] env[61868]: DEBUG nova.virt.hardware [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 720.039917] env[61868]: DEBUG nova.virt.hardware [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 720.040110] env[61868]: DEBUG nova.virt.hardware [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 720.041295] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaba4460-6a65-446b-8b08-88045ccb705c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.051474] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f59da77-cef3-4094-80cd-082b7b0c7f89 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.069342] env[61868]: ERROR nova.compute.manager [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9fe37d4c-b367-48ea-b471-3a851ec07246, please check neutron logs for more information. [ 720.069342] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Traceback (most recent call last): [ 720.069342] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 720.069342] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] yield resources [ 720.069342] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 720.069342] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] self.driver.spawn(context, instance, image_meta, [ 720.069342] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 720.069342] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 720.069342] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 720.069342] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] vm_ref = self.build_virtual_machine(instance, [ 720.069342] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 720.069834] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] vif_infos = vmwarevif.get_vif_info(self._session, [ 720.069834] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 720.069834] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] for vif in network_info: [ 720.069834] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 720.069834] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] return self._sync_wrapper(fn, *args, **kwargs) [ 720.069834] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 720.069834] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] self.wait() [ 720.069834] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 720.069834] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] self[:] = self._gt.wait() [ 720.069834] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 720.069834] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] return self._exit_event.wait() [ 720.069834] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 720.069834] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] current.throw(*self._exc) [ 720.070247] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 720.070247] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] result = function(*args, **kwargs) [ 720.070247] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 720.070247] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] return func(*args, **kwargs) [ 720.070247] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 720.070247] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] raise e [ 720.070247] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 720.070247] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] nwinfo = self.network_api.allocate_for_instance( [ 720.070247] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 720.070247] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] created_port_ids = self._update_ports_for_instance( [ 720.070247] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 720.070247] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] with excutils.save_and_reraise_exception(): [ 720.070247] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 720.070682] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] self.force_reraise() [ 720.070682] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 720.070682] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] raise self.value [ 720.070682] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 720.070682] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] updated_port = self._update_port( [ 720.070682] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 720.070682] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] _ensure_no_port_binding_failure(port) [ 720.070682] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 720.070682] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] raise exception.PortBindingFailed(port_id=port['id']) [ 720.070682] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] nova.exception.PortBindingFailed: Binding failed for port 9fe37d4c-b367-48ea-b471-3a851ec07246, please check neutron logs for more information. [ 720.070682] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] [ 720.070682] env[61868]: INFO nova.compute.manager [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Terminating instance [ 720.071897] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "refresh_cache-30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.072064] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquired lock "refresh_cache-30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.072229] env[61868]: DEBUG nova.network.neutron [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 720.283597] env[61868]: DEBUG nova.compute.manager [req-a2cac58b-f4d8-4bbe-b98e-17a8986c841f req-c69195cb-896b-4257-b10a-96381b8cab6a service nova] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Received event network-changed-9fe37d4c-b367-48ea-b471-3a851ec07246 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 720.283597] env[61868]: DEBUG nova.compute.manager [req-a2cac58b-f4d8-4bbe-b98e-17a8986c841f req-c69195cb-896b-4257-b10a-96381b8cab6a service nova] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Refreshing instance network info cache due to event network-changed-9fe37d4c-b367-48ea-b471-3a851ec07246. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 720.283751] env[61868]: DEBUG oslo_concurrency.lockutils [req-a2cac58b-f4d8-4bbe-b98e-17a8986c841f req-c69195cb-896b-4257-b10a-96381b8cab6a service nova] Acquiring lock "refresh_cache-30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.421868] env[61868]: DEBUG oslo_concurrency.lockutils [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.432s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.422423] env[61868]: DEBUG nova.compute.manager [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 720.425028] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.585s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.426677] env[61868]: INFO nova.compute.claims [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 720.589515] env[61868]: DEBUG nova.network.neutron [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.670184] env[61868]: DEBUG nova.network.neutron [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.936325] env[61868]: DEBUG nova.compute.utils [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 720.939688] env[61868]: DEBUG nova.compute.manager [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 720.939862] env[61868]: DEBUG nova.network.neutron [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 720.981844] env[61868]: DEBUG nova.policy [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '65f6dc7b309f493f97dbc41852ca2a91', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '235d12c7819640b483f7c061fe103609', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 721.173096] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Releasing lock "refresh_cache-30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.173558] env[61868]: DEBUG nova.compute.manager [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 721.173832] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 721.174111] env[61868]: DEBUG oslo_concurrency.lockutils [req-a2cac58b-f4d8-4bbe-b98e-17a8986c841f req-c69195cb-896b-4257-b10a-96381b8cab6a service nova] Acquired lock "refresh_cache-30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.174290] env[61868]: DEBUG nova.network.neutron [req-a2cac58b-f4d8-4bbe-b98e-17a8986c841f req-c69195cb-896b-4257-b10a-96381b8cab6a service nova] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Refreshing network info cache for port 9fe37d4c-b367-48ea-b471-3a851ec07246 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 721.176802] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ae000127-a8c4-40f0-85c2-01f57310ced4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.184508] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31051649-771c-4d22-b7c7-47a19b27aded {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.207931] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5 could not be found. [ 721.208177] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 721.208362] env[61868]: INFO nova.compute.manager [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Took 0.03 seconds to destroy the instance on the hypervisor. [ 721.208600] env[61868]: DEBUG oslo.service.loopingcall [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 721.208842] env[61868]: DEBUG nova.compute.manager [-] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 721.208937] env[61868]: DEBUG nova.network.neutron [-] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 721.226504] env[61868]: DEBUG nova.network.neutron [-] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.240152] env[61868]: DEBUG nova.network.neutron [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Successfully created port: 5a36ed2f-6282-43a3-840f-9ab052883ad3 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 721.440997] env[61868]: DEBUG nova.compute.manager [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 721.465419] env[61868]: DEBUG nova.network.neutron [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Successfully created port: f99d952b-4b48-4143-b974-baeea5ecf060 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 721.699682] env[61868]: DEBUG nova.network.neutron [req-a2cac58b-f4d8-4bbe-b98e-17a8986c841f req-c69195cb-896b-4257-b10a-96381b8cab6a service nova] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.728627] env[61868]: DEBUG nova.network.neutron [-] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.824154] env[61868]: DEBUG nova.network.neutron [req-a2cac58b-f4d8-4bbe-b98e-17a8986c841f req-c69195cb-896b-4257-b10a-96381b8cab6a service nova] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.841881] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af9a85ce-2d35-47ef-82c4-e97c5d2b89d2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.849938] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6216aebf-e9fe-447f-bf99-95082c596b34 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.880532] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b451fb4e-f955-4f5f-a019-63e4f81f4f2a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.887824] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48b6de20-d032-4d12-bd23-c7d5c2944fc1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.901013] env[61868]: DEBUG nova.compute.provider_tree [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 722.232733] env[61868]: INFO nova.compute.manager [-] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Took 1.02 seconds to deallocate network for instance. [ 722.235107] env[61868]: DEBUG nova.compute.claims [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 722.235288] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.314060] env[61868]: DEBUG nova.compute.manager [req-ed52b302-6ecd-4ddf-a0d4-961619f9cb61 req-db7728eb-e341-492e-9131-c117f27922c0 service nova] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Received event network-changed-5a36ed2f-6282-43a3-840f-9ab052883ad3 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 722.314267] env[61868]: DEBUG nova.compute.manager [req-ed52b302-6ecd-4ddf-a0d4-961619f9cb61 req-db7728eb-e341-492e-9131-c117f27922c0 service nova] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Refreshing instance network info cache due to event network-changed-5a36ed2f-6282-43a3-840f-9ab052883ad3. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 722.314479] env[61868]: DEBUG oslo_concurrency.lockutils [req-ed52b302-6ecd-4ddf-a0d4-961619f9cb61 req-db7728eb-e341-492e-9131-c117f27922c0 service nova] Acquiring lock "refresh_cache-b4a7ff58-0f48-4942-a270-de6910ff0c0c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.314620] env[61868]: DEBUG oslo_concurrency.lockutils [req-ed52b302-6ecd-4ddf-a0d4-961619f9cb61 req-db7728eb-e341-492e-9131-c117f27922c0 service nova] Acquired lock "refresh_cache-b4a7ff58-0f48-4942-a270-de6910ff0c0c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.314777] env[61868]: DEBUG nova.network.neutron [req-ed52b302-6ecd-4ddf-a0d4-961619f9cb61 req-db7728eb-e341-492e-9131-c117f27922c0 service nova] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Refreshing network info cache for port 5a36ed2f-6282-43a3-840f-9ab052883ad3 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 722.327073] env[61868]: DEBUG oslo_concurrency.lockutils [req-a2cac58b-f4d8-4bbe-b98e-17a8986c841f req-c69195cb-896b-4257-b10a-96381b8cab6a service nova] Releasing lock "refresh_cache-30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.327443] env[61868]: DEBUG nova.compute.manager [req-a2cac58b-f4d8-4bbe-b98e-17a8986c841f req-c69195cb-896b-4257-b10a-96381b8cab6a service nova] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Received event network-vif-deleted-9fe37d4c-b367-48ea-b471-3a851ec07246 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 722.406044] env[61868]: DEBUG nova.scheduler.client.report [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 722.432847] env[61868]: ERROR nova.compute.manager [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5a36ed2f-6282-43a3-840f-9ab052883ad3, please check neutron logs for more information. [ 722.432847] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 722.432847] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 722.432847] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 722.432847] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 722.432847] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 722.432847] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 722.432847] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 722.432847] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 722.432847] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 722.432847] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 722.432847] env[61868]: ERROR nova.compute.manager raise self.value [ 722.432847] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 722.432847] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 722.432847] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 722.432847] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 722.433315] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 722.433315] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 722.433315] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5a36ed2f-6282-43a3-840f-9ab052883ad3, please check neutron logs for more information. [ 722.433315] env[61868]: ERROR nova.compute.manager [ 722.433315] env[61868]: Traceback (most recent call last): [ 722.433315] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 722.433315] env[61868]: listener.cb(fileno) [ 722.433315] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 722.433315] env[61868]: result = function(*args, **kwargs) [ 722.433315] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 722.433315] env[61868]: return func(*args, **kwargs) [ 722.433315] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 722.433315] env[61868]: raise e [ 722.433315] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 722.433315] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 722.433315] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 722.433315] env[61868]: created_port_ids = self._update_ports_for_instance( [ 722.433315] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 722.433315] env[61868]: with excutils.save_and_reraise_exception(): [ 722.433315] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 722.433315] env[61868]: self.force_reraise() [ 722.433315] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 722.433315] env[61868]: raise self.value [ 722.433315] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 722.433315] env[61868]: updated_port = self._update_port( [ 722.433315] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 722.433315] env[61868]: _ensure_no_port_binding_failure(port) [ 722.433315] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 722.433315] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 722.434148] env[61868]: nova.exception.PortBindingFailed: Binding failed for port 5a36ed2f-6282-43a3-840f-9ab052883ad3, please check neutron logs for more information. [ 722.434148] env[61868]: Removing descriptor: 16 [ 722.454039] env[61868]: DEBUG nova.compute.manager [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 722.477320] env[61868]: DEBUG nova.virt.hardware [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 722.477565] env[61868]: DEBUG nova.virt.hardware [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 722.478072] env[61868]: DEBUG nova.virt.hardware [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 722.478072] env[61868]: DEBUG nova.virt.hardware [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 722.478072] env[61868]: DEBUG nova.virt.hardware [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 722.478225] env[61868]: DEBUG nova.virt.hardware [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 722.478392] env[61868]: DEBUG nova.virt.hardware [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 722.478550] env[61868]: DEBUG nova.virt.hardware [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 722.478860] env[61868]: DEBUG nova.virt.hardware [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 722.479099] env[61868]: DEBUG nova.virt.hardware [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 722.479294] env[61868]: DEBUG nova.virt.hardware [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 722.480151] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a10e7d33-0b14-44f5-96b2-e7419d3939af {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.489784] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f54ae86-ae7b-4726-9e87-cfff1612f56a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.503284] env[61868]: ERROR nova.compute.manager [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5a36ed2f-6282-43a3-840f-9ab052883ad3, please check neutron logs for more information. [ 722.503284] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Traceback (most recent call last): [ 722.503284] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 722.503284] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] yield resources [ 722.503284] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 722.503284] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] self.driver.spawn(context, instance, image_meta, [ 722.503284] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 722.503284] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 722.503284] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 722.503284] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] vm_ref = self.build_virtual_machine(instance, [ 722.503284] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 722.503855] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] vif_infos = vmwarevif.get_vif_info(self._session, [ 722.503855] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 722.503855] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] for vif in network_info: [ 722.503855] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 722.503855] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] return self._sync_wrapper(fn, *args, **kwargs) [ 722.503855] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 722.503855] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] self.wait() [ 722.503855] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 722.503855] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] self[:] = self._gt.wait() [ 722.503855] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 722.503855] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] return self._exit_event.wait() [ 722.503855] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 722.503855] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] current.throw(*self._exc) [ 722.504288] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 722.504288] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] result = function(*args, **kwargs) [ 722.504288] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 722.504288] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] return func(*args, **kwargs) [ 722.504288] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 722.504288] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] raise e [ 722.504288] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 722.504288] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] nwinfo = self.network_api.allocate_for_instance( [ 722.504288] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 722.504288] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] created_port_ids = self._update_ports_for_instance( [ 722.504288] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 722.504288] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] with excutils.save_and_reraise_exception(): [ 722.504288] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 722.504612] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] self.force_reraise() [ 722.504612] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 722.504612] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] raise self.value [ 722.504612] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 722.504612] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] updated_port = self._update_port( [ 722.504612] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 722.504612] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] _ensure_no_port_binding_failure(port) [ 722.504612] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 722.504612] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] raise exception.PortBindingFailed(port_id=port['id']) [ 722.504612] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] nova.exception.PortBindingFailed: Binding failed for port 5a36ed2f-6282-43a3-840f-9ab052883ad3, please check neutron logs for more information. [ 722.504612] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] [ 722.504612] env[61868]: INFO nova.compute.manager [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Terminating instance [ 722.505525] env[61868]: DEBUG oslo_concurrency.lockutils [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Acquiring lock "refresh_cache-b4a7ff58-0f48-4942-a270-de6910ff0c0c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.843953] env[61868]: DEBUG nova.network.neutron [req-ed52b302-6ecd-4ddf-a0d4-961619f9cb61 req-db7728eb-e341-492e-9131-c117f27922c0 service nova] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 722.909936] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.485s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.910478] env[61868]: DEBUG nova.compute.manager [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 722.913156] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.665s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 722.930610] env[61868]: DEBUG nova.network.neutron [req-ed52b302-6ecd-4ddf-a0d4-961619f9cb61 req-db7728eb-e341-492e-9131-c117f27922c0 service nova] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.418072] env[61868]: DEBUG nova.compute.utils [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 723.422690] env[61868]: DEBUG nova.compute.manager [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 723.422690] env[61868]: DEBUG nova.network.neutron [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 723.432090] env[61868]: DEBUG oslo_concurrency.lockutils [req-ed52b302-6ecd-4ddf-a0d4-961619f9cb61 req-db7728eb-e341-492e-9131-c117f27922c0 service nova] Releasing lock "refresh_cache-b4a7ff58-0f48-4942-a270-de6910ff0c0c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.432667] env[61868]: DEBUG oslo_concurrency.lockutils [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Acquired lock "refresh_cache-b4a7ff58-0f48-4942-a270-de6910ff0c0c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.432855] env[61868]: DEBUG nova.network.neutron [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 723.485547] env[61868]: DEBUG nova.policy [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9a68749e736f43c3a686e16b85eb964b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6f592ea4ad384a8f9669a09eb3ed9bb2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 723.753937] env[61868]: DEBUG nova.network.neutron [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Successfully created port: 9967a7d8-aea8-47aa-b6e5-2d7eb28a285a {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 723.788428] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eee088c3-88f3-4a99-97a0-d58db4c84e6f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.796924] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6604cc7c-5f52-408f-a50c-5d0a5330bd73 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.830803] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f150ced-d9f3-4a20-9ed0-befdb0767ca9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.838663] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe93c5a9-5dff-4eaf-b3ba-129e3a86ca42 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.854203] env[61868]: DEBUG nova.compute.provider_tree [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 723.925712] env[61868]: DEBUG nova.compute.manager [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 723.985571] env[61868]: DEBUG nova.network.neutron [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 724.130881] env[61868]: DEBUG nova.network.neutron [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.344693] env[61868]: DEBUG nova.compute.manager [req-11be88b8-307c-4947-8300-253d926361a0 req-5d1d4800-bd9b-434a-8f49-fc06085bd742 service nova] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Received event network-vif-deleted-5a36ed2f-6282-43a3-840f-9ab052883ad3 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 724.359892] env[61868]: DEBUG nova.scheduler.client.report [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 724.633392] env[61868]: DEBUG oslo_concurrency.lockutils [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Releasing lock "refresh_cache-b4a7ff58-0f48-4942-a270-de6910ff0c0c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 724.633852] env[61868]: DEBUG nova.compute.manager [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 724.634093] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 724.634412] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-01815e7a-e915-4918-8c84-541b209505f0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.643537] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef4095a9-6efc-49e8-ae75-a6dac0268423 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.665436] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b4a7ff58-0f48-4942-a270-de6910ff0c0c could not be found. [ 724.665673] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 724.665992] env[61868]: INFO nova.compute.manager [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Took 0.03 seconds to destroy the instance on the hypervisor. [ 724.666101] env[61868]: DEBUG oslo.service.loopingcall [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 724.666335] env[61868]: DEBUG nova.compute.manager [-] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 724.666503] env[61868]: DEBUG nova.network.neutron [-] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 724.708992] env[61868]: ERROR nova.compute.manager [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9967a7d8-aea8-47aa-b6e5-2d7eb28a285a, please check neutron logs for more information. [ 724.708992] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 724.708992] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 724.708992] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 724.708992] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 724.708992] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 724.708992] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 724.708992] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 724.708992] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.708992] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 724.708992] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.708992] env[61868]: ERROR nova.compute.manager raise self.value [ 724.708992] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 724.708992] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 724.708992] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.708992] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 724.709429] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.709429] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 724.709429] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9967a7d8-aea8-47aa-b6e5-2d7eb28a285a, please check neutron logs for more information. [ 724.709429] env[61868]: ERROR nova.compute.manager [ 724.709429] env[61868]: Traceback (most recent call last): [ 724.709429] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 724.709429] env[61868]: listener.cb(fileno) [ 724.709429] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 724.709429] env[61868]: result = function(*args, **kwargs) [ 724.709429] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 724.709429] env[61868]: return func(*args, **kwargs) [ 724.709429] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 724.709429] env[61868]: raise e [ 724.709429] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 724.709429] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 724.709429] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 724.709429] env[61868]: created_port_ids = self._update_ports_for_instance( [ 724.709429] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 724.709429] env[61868]: with excutils.save_and_reraise_exception(): [ 724.709429] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.709429] env[61868]: self.force_reraise() [ 724.709429] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.709429] env[61868]: raise self.value [ 724.709429] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 724.709429] env[61868]: updated_port = self._update_port( [ 724.709429] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.709429] env[61868]: _ensure_no_port_binding_failure(port) [ 724.709429] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.709429] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 724.710163] env[61868]: nova.exception.PortBindingFailed: Binding failed for port 9967a7d8-aea8-47aa-b6e5-2d7eb28a285a, please check neutron logs for more information. [ 724.710163] env[61868]: Removing descriptor: 16 [ 724.745441] env[61868]: DEBUG nova.network.neutron [-] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 724.863956] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.951s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.865170] env[61868]: ERROR nova.compute.manager [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 48d4c453-40b0-4fb2-a83c-b145b3e20694, please check neutron logs for more information. [ 724.865170] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Traceback (most recent call last): [ 724.865170] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 724.865170] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] self.driver.spawn(context, instance, image_meta, [ 724.865170] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 724.865170] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 724.865170] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 724.865170] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] vm_ref = self.build_virtual_machine(instance, [ 724.865170] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 724.865170] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] vif_infos = vmwarevif.get_vif_info(self._session, [ 724.865170] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 724.866397] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] for vif in network_info: [ 724.866397] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 724.866397] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] return self._sync_wrapper(fn, *args, **kwargs) [ 724.866397] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 724.866397] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] self.wait() [ 724.866397] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 724.866397] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] self[:] = self._gt.wait() [ 724.866397] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 724.866397] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] return self._exit_event.wait() [ 724.866397] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 724.866397] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] current.throw(*self._exc) [ 724.866397] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 724.866397] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] result = function(*args, **kwargs) [ 724.867223] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 724.867223] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] return func(*args, **kwargs) [ 724.867223] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 724.867223] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] raise e [ 724.867223] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 724.867223] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] nwinfo = self.network_api.allocate_for_instance( [ 724.867223] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 724.867223] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] created_port_ids = self._update_ports_for_instance( [ 724.867223] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 724.867223] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] with excutils.save_and_reraise_exception(): [ 724.867223] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.867223] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] self.force_reraise() [ 724.867223] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.867777] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] raise self.value [ 724.867777] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 724.867777] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] updated_port = self._update_port( [ 724.867777] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.867777] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] _ensure_no_port_binding_failure(port) [ 724.867777] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.867777] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] raise exception.PortBindingFailed(port_id=port['id']) [ 724.867777] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] nova.exception.PortBindingFailed: Binding failed for port 48d4c453-40b0-4fb2-a83c-b145b3e20694, please check neutron logs for more information. [ 724.867777] env[61868]: ERROR nova.compute.manager [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] [ 724.867777] env[61868]: DEBUG nova.compute.utils [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Binding failed for port 48d4c453-40b0-4fb2-a83c-b145b3e20694, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 724.868214] env[61868]: DEBUG nova.compute.manager [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Build of instance c0c9e61b-ab2b-47dc-997b-570204ab4c3b was re-scheduled: Binding failed for port 48d4c453-40b0-4fb2-a83c-b145b3e20694, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 724.868214] env[61868]: DEBUG nova.compute.manager [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 724.868214] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] Acquiring lock "refresh_cache-c0c9e61b-ab2b-47dc-997b-570204ab4c3b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.868214] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] Acquired lock "refresh_cache-c0c9e61b-ab2b-47dc-997b-570204ab4c3b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.868468] env[61868]: DEBUG nova.network.neutron [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 724.869421] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.825s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.871166] env[61868]: INFO nova.compute.claims [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 724.934933] env[61868]: DEBUG nova.compute.manager [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 724.963542] env[61868]: DEBUG nova.virt.hardware [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:20:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='b9bd276b-d827-441d-a42e-2ac6458e8961',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-459096521',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 724.964413] env[61868]: DEBUG nova.virt.hardware [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 724.964413] env[61868]: DEBUG nova.virt.hardware [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 724.964413] env[61868]: DEBUG nova.virt.hardware [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 724.964413] env[61868]: DEBUG nova.virt.hardware [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 724.964631] env[61868]: DEBUG nova.virt.hardware [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 724.964631] env[61868]: DEBUG nova.virt.hardware [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 724.964867] env[61868]: DEBUG nova.virt.hardware [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 724.964932] env[61868]: DEBUG nova.virt.hardware [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 724.965856] env[61868]: DEBUG nova.virt.hardware [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 724.965856] env[61868]: DEBUG nova.virt.hardware [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 724.966226] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e6038da-42d4-4ee7-b7b0-22ab2c7cb883 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.974661] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67db23a6-70af-45ec-b2cb-8c5d1139fe18 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.990412] env[61868]: ERROR nova.compute.manager [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9967a7d8-aea8-47aa-b6e5-2d7eb28a285a, please check neutron logs for more information. [ 724.990412] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Traceback (most recent call last): [ 724.990412] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 724.990412] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] yield resources [ 724.990412] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 724.990412] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] self.driver.spawn(context, instance, image_meta, [ 724.990412] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 724.990412] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 724.990412] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 724.990412] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] vm_ref = self.build_virtual_machine(instance, [ 724.990412] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 724.990816] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] vif_infos = vmwarevif.get_vif_info(self._session, [ 724.990816] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 724.990816] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] for vif in network_info: [ 724.990816] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 724.990816] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] return self._sync_wrapper(fn, *args, **kwargs) [ 724.990816] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 724.990816] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] self.wait() [ 724.990816] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 724.990816] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] self[:] = self._gt.wait() [ 724.990816] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 724.990816] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] return self._exit_event.wait() [ 724.990816] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 724.990816] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] current.throw(*self._exc) [ 724.991188] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 724.991188] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] result = function(*args, **kwargs) [ 724.991188] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 724.991188] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] return func(*args, **kwargs) [ 724.991188] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 724.991188] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] raise e [ 724.991188] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 724.991188] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] nwinfo = self.network_api.allocate_for_instance( [ 724.991188] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 724.991188] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] created_port_ids = self._update_ports_for_instance( [ 724.991188] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 724.991188] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] with excutils.save_and_reraise_exception(): [ 724.991188] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.991545] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] self.force_reraise() [ 724.991545] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.991545] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] raise self.value [ 724.991545] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 724.991545] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] updated_port = self._update_port( [ 724.991545] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.991545] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] _ensure_no_port_binding_failure(port) [ 724.991545] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.991545] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] raise exception.PortBindingFailed(port_id=port['id']) [ 724.991545] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] nova.exception.PortBindingFailed: Binding failed for port 9967a7d8-aea8-47aa-b6e5-2d7eb28a285a, please check neutron logs for more information. [ 724.991545] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] [ 724.991545] env[61868]: INFO nova.compute.manager [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Terminating instance [ 724.993535] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Acquiring lock "refresh_cache-92d3557c-655f-45d9-ae7d-bbe62567e1e2" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.993738] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Acquired lock "refresh_cache-92d3557c-655f-45d9-ae7d-bbe62567e1e2" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.993945] env[61868]: DEBUG nova.network.neutron [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 725.387542] env[61868]: DEBUG nova.network.neutron [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 725.430668] env[61868]: DEBUG nova.network.neutron [-] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.470511] env[61868]: DEBUG nova.network.neutron [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.509227] env[61868]: DEBUG nova.network.neutron [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 725.554042] env[61868]: DEBUG nova.network.neutron [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.933374] env[61868]: INFO nova.compute.manager [-] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Took 1.27 seconds to deallocate network for instance. [ 725.936023] env[61868]: DEBUG nova.compute.claims [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 725.936289] env[61868]: DEBUG oslo_concurrency.lockutils [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.973086] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] Releasing lock "refresh_cache-c0c9e61b-ab2b-47dc-997b-570204ab4c3b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.973284] env[61868]: DEBUG nova.compute.manager [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 725.973534] env[61868]: DEBUG nova.compute.manager [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 725.973703] env[61868]: DEBUG nova.network.neutron [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 725.989584] env[61868]: DEBUG nova.network.neutron [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.056056] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Releasing lock "refresh_cache-92d3557c-655f-45d9-ae7d-bbe62567e1e2" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.056452] env[61868]: DEBUG nova.compute.manager [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 726.056645] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 726.056951] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7d697191-7bc2-4737-b31a-4ceb1bf9ee5f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.067371] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26133476-dc0e-409e-b9c0-981c73c5e732 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.091430] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 92d3557c-655f-45d9-ae7d-bbe62567e1e2 could not be found. [ 726.091663] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 726.091866] env[61868]: INFO nova.compute.manager [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Took 0.04 seconds to destroy the instance on the hypervisor. [ 726.092128] env[61868]: DEBUG oslo.service.loopingcall [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 726.094424] env[61868]: DEBUG nova.compute.manager [-] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 726.094523] env[61868]: DEBUG nova.network.neutron [-] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 726.111018] env[61868]: DEBUG nova.network.neutron [-] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.207130] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa72b2a3-7351-4de6-8205-b5c6810df0b6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.214391] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d138a68c-793c-486f-976f-0c8c02cacf7a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.245088] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd77e6ba-d00d-4e44-96d3-441d259f8b72 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.252760] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0b5b17b-5c4d-4fd2-890b-c8cba1777ed1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.266032] env[61868]: DEBUG nova.compute.provider_tree [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 726.390202] env[61868]: DEBUG nova.compute.manager [req-8dfcd50b-0449-4d6e-b122-fa08b03fd44e req-ce4ab7a2-7ed5-4846-82f5-bd12a3f8278c service nova] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Received event network-changed-9967a7d8-aea8-47aa-b6e5-2d7eb28a285a {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 726.390380] env[61868]: DEBUG nova.compute.manager [req-8dfcd50b-0449-4d6e-b122-fa08b03fd44e req-ce4ab7a2-7ed5-4846-82f5-bd12a3f8278c service nova] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Refreshing instance network info cache due to event network-changed-9967a7d8-aea8-47aa-b6e5-2d7eb28a285a. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 726.390589] env[61868]: DEBUG oslo_concurrency.lockutils [req-8dfcd50b-0449-4d6e-b122-fa08b03fd44e req-ce4ab7a2-7ed5-4846-82f5-bd12a3f8278c service nova] Acquiring lock "refresh_cache-92d3557c-655f-45d9-ae7d-bbe62567e1e2" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.390790] env[61868]: DEBUG oslo_concurrency.lockutils [req-8dfcd50b-0449-4d6e-b122-fa08b03fd44e req-ce4ab7a2-7ed5-4846-82f5-bd12a3f8278c service nova] Acquired lock "refresh_cache-92d3557c-655f-45d9-ae7d-bbe62567e1e2" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.390972] env[61868]: DEBUG nova.network.neutron [req-8dfcd50b-0449-4d6e-b122-fa08b03fd44e req-ce4ab7a2-7ed5-4846-82f5-bd12a3f8278c service nova] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Refreshing network info cache for port 9967a7d8-aea8-47aa-b6e5-2d7eb28a285a {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 726.493689] env[61868]: DEBUG nova.network.neutron [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.611328] env[61868]: DEBUG nova.network.neutron [-] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.768880] env[61868]: DEBUG nova.scheduler.client.report [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 726.909227] env[61868]: DEBUG nova.network.neutron [req-8dfcd50b-0449-4d6e-b122-fa08b03fd44e req-ce4ab7a2-7ed5-4846-82f5-bd12a3f8278c service nova] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.955082] env[61868]: DEBUG nova.network.neutron [req-8dfcd50b-0449-4d6e-b122-fa08b03fd44e req-ce4ab7a2-7ed5-4846-82f5-bd12a3f8278c service nova] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.996730] env[61868]: INFO nova.compute.manager [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] [instance: c0c9e61b-ab2b-47dc-997b-570204ab4c3b] Took 1.02 seconds to deallocate network for instance. [ 727.114926] env[61868]: INFO nova.compute.manager [-] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Took 1.02 seconds to deallocate network for instance. [ 727.117602] env[61868]: DEBUG nova.compute.claims [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 727.117602] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.273383] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.404s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.273665] env[61868]: DEBUG nova.compute.manager [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 727.276402] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.733s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 727.457840] env[61868]: DEBUG oslo_concurrency.lockutils [req-8dfcd50b-0449-4d6e-b122-fa08b03fd44e req-ce4ab7a2-7ed5-4846-82f5-bd12a3f8278c service nova] Releasing lock "refresh_cache-92d3557c-655f-45d9-ae7d-bbe62567e1e2" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.458137] env[61868]: DEBUG nova.compute.manager [req-8dfcd50b-0449-4d6e-b122-fa08b03fd44e req-ce4ab7a2-7ed5-4846-82f5-bd12a3f8278c service nova] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Received event network-vif-deleted-9967a7d8-aea8-47aa-b6e5-2d7eb28a285a {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 727.781120] env[61868]: DEBUG nova.compute.utils [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 727.782804] env[61868]: DEBUG nova.compute.manager [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 727.782982] env[61868]: DEBUG nova.network.neutron [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 727.823312] env[61868]: DEBUG nova.policy [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1bf93172b7584bc19a2bf06ee6479870', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0b09de1886944903a10f07f07a0362d6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 728.023380] env[61868]: INFO nova.scheduler.client.report [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] Deleted allocations for instance c0c9e61b-ab2b-47dc-997b-570204ab4c3b [ 728.111227] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f8d3217-d18f-420d-87c3-f27f8419206e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.121345] env[61868]: DEBUG nova.network.neutron [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Successfully created port: 4c4da9ff-f957-4820-8c63-da666cca0fa6 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 728.123413] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d64ef11-89ad-47a0-91eb-e2940bc04322 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.157122] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9a6ebce-a985-48de-bbd8-0d2eb3abb998 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.165443] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fe98bdf-76bc-4e8e-807d-0bb47d10d274 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.180277] env[61868]: DEBUG nova.compute.provider_tree [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 728.288395] env[61868]: DEBUG nova.compute.manager [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 728.532307] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ab17370c-4672-47d6-994c-8b86635d2af2 tempest-ServersTestManualDisk-763917599 tempest-ServersTestManualDisk-763917599-project-member] Lock "c0c9e61b-ab2b-47dc-997b-570204ab4c3b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 143.111s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.683665] env[61868]: DEBUG nova.scheduler.client.report [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 728.794253] env[61868]: INFO nova.virt.block_device [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Booting with volume e5ff56ef-31a8-40b2-ac62-9668a4116b23 at /dev/sda [ 728.814419] env[61868]: DEBUG nova.compute.manager [req-57f309d2-f9d8-4439-a8f4-fe0501fec8f0 req-b82693df-eb27-4c9c-bc49-73fd95b55a56 service nova] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Received event network-changed-4c4da9ff-f957-4820-8c63-da666cca0fa6 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 728.814613] env[61868]: DEBUG nova.compute.manager [req-57f309d2-f9d8-4439-a8f4-fe0501fec8f0 req-b82693df-eb27-4c9c-bc49-73fd95b55a56 service nova] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Refreshing instance network info cache due to event network-changed-4c4da9ff-f957-4820-8c63-da666cca0fa6. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 728.816105] env[61868]: DEBUG oslo_concurrency.lockutils [req-57f309d2-f9d8-4439-a8f4-fe0501fec8f0 req-b82693df-eb27-4c9c-bc49-73fd95b55a56 service nova] Acquiring lock "refresh_cache-66c5f446-91e5-4de1-abb6-f06a0a36020d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.816105] env[61868]: DEBUG oslo_concurrency.lockutils [req-57f309d2-f9d8-4439-a8f4-fe0501fec8f0 req-b82693df-eb27-4c9c-bc49-73fd95b55a56 service nova] Acquired lock "refresh_cache-66c5f446-91e5-4de1-abb6-f06a0a36020d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.816105] env[61868]: DEBUG nova.network.neutron [req-57f309d2-f9d8-4439-a8f4-fe0501fec8f0 req-b82693df-eb27-4c9c-bc49-73fd95b55a56 service nova] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Refreshing network info cache for port 4c4da9ff-f957-4820-8c63-da666cca0fa6 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 728.850096] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b179ba62-23ec-48e4-8c36-d03c89e9d67c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.860790] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45674592-9b7c-464c-be85-970fdaaea229 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.883792] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-807f1fb3-daaf-4d90-849c-e3085ae03abe {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.896740] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd1eec3d-2ea5-4dd1-856b-a81faca63b3d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.919811] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95c8651e-ee58-4c6f-b5d8-15612aa85fc5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.925772] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ca6cff1-ba70-4033-b279-183a17a7ac47 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.940050] env[61868]: DEBUG nova.virt.block_device [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Updating existing volume attachment record: aa060854-ce4f-41fe-b70f-9c0a4c98216a {{(pid=61868) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 729.013270] env[61868]: ERROR nova.compute.manager [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4c4da9ff-f957-4820-8c63-da666cca0fa6, please check neutron logs for more information. [ 729.013270] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 729.013270] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 729.013270] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 729.013270] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 729.013270] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 729.013270] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 729.013270] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 729.013270] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 729.013270] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 729.013270] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 729.013270] env[61868]: ERROR nova.compute.manager raise self.value [ 729.013270] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 729.013270] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 729.013270] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 729.013270] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 729.013734] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 729.013734] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 729.013734] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4c4da9ff-f957-4820-8c63-da666cca0fa6, please check neutron logs for more information. [ 729.013734] env[61868]: ERROR nova.compute.manager [ 729.013734] env[61868]: Traceback (most recent call last): [ 729.013734] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 729.013734] env[61868]: listener.cb(fileno) [ 729.013925] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 729.013925] env[61868]: result = function(*args, **kwargs) [ 729.013925] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 729.013925] env[61868]: return func(*args, **kwargs) [ 729.013925] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 729.013925] env[61868]: raise e [ 729.013925] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 729.013925] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 729.013925] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 729.013925] env[61868]: created_port_ids = self._update_ports_for_instance( [ 729.013925] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 729.013925] env[61868]: with excutils.save_and_reraise_exception(): [ 729.013925] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 729.013925] env[61868]: self.force_reraise() [ 729.013925] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 729.013925] env[61868]: raise self.value [ 729.013925] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 729.013925] env[61868]: updated_port = self._update_port( [ 729.013925] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 729.013925] env[61868]: _ensure_no_port_binding_failure(port) [ 729.013925] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 729.013925] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 729.013925] env[61868]: nova.exception.PortBindingFailed: Binding failed for port 4c4da9ff-f957-4820-8c63-da666cca0fa6, please check neutron logs for more information. [ 729.013925] env[61868]: Removing descriptor: 15 [ 729.035587] env[61868]: DEBUG nova.compute.manager [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 729.191789] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.915s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.192501] env[61868]: ERROR nova.compute.manager [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 77c9c449-01ba-4bab-90f8-49033ffd0a67, please check neutron logs for more information. [ 729.192501] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Traceback (most recent call last): [ 729.192501] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 729.192501] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] self.driver.spawn(context, instance, image_meta, [ 729.192501] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 729.192501] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] self._vmops.spawn(context, instance, image_meta, injected_files, [ 729.192501] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 729.192501] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] vm_ref = self.build_virtual_machine(instance, [ 729.192501] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 729.192501] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] vif_infos = vmwarevif.get_vif_info(self._session, [ 729.192501] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 729.192860] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] for vif in network_info: [ 729.192860] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 729.192860] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] return self._sync_wrapper(fn, *args, **kwargs) [ 729.192860] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 729.192860] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] self.wait() [ 729.192860] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 729.192860] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] self[:] = self._gt.wait() [ 729.192860] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 729.192860] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] return self._exit_event.wait() [ 729.192860] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 729.192860] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] current.throw(*self._exc) [ 729.192860] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 729.192860] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] result = function(*args, **kwargs) [ 729.193227] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 729.193227] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] return func(*args, **kwargs) [ 729.193227] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 729.193227] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] raise e [ 729.193227] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 729.193227] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] nwinfo = self.network_api.allocate_for_instance( [ 729.193227] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 729.193227] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] created_port_ids = self._update_ports_for_instance( [ 729.193227] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 729.193227] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] with excutils.save_and_reraise_exception(): [ 729.193227] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 729.193227] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] self.force_reraise() [ 729.193227] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 729.193548] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] raise self.value [ 729.193548] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 729.193548] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] updated_port = self._update_port( [ 729.193548] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 729.193548] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] _ensure_no_port_binding_failure(port) [ 729.193548] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 729.193548] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] raise exception.PortBindingFailed(port_id=port['id']) [ 729.193548] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] nova.exception.PortBindingFailed: Binding failed for port 77c9c449-01ba-4bab-90f8-49033ffd0a67, please check neutron logs for more information. [ 729.193548] env[61868]: ERROR nova.compute.manager [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] [ 729.193548] env[61868]: DEBUG nova.compute.utils [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Binding failed for port 77c9c449-01ba-4bab-90f8-49033ffd0a67, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 729.194741] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.847s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.196429] env[61868]: INFO nova.compute.claims [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 729.199536] env[61868]: DEBUG nova.compute.manager [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Build of instance 96694bb9-fbf2-4b71-9cb5-37f81b634992 was re-scheduled: Binding failed for port 77c9c449-01ba-4bab-90f8-49033ffd0a67, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 729.200043] env[61868]: DEBUG nova.compute.manager [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 729.200402] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquiring lock "refresh_cache-96694bb9-fbf2-4b71-9cb5-37f81b634992" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.200616] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquired lock "refresh_cache-96694bb9-fbf2-4b71-9cb5-37f81b634992" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.200953] env[61868]: DEBUG nova.network.neutron [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 729.334943] env[61868]: DEBUG nova.network.neutron [req-57f309d2-f9d8-4439-a8f4-fe0501fec8f0 req-b82693df-eb27-4c9c-bc49-73fd95b55a56 service nova] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 729.450248] env[61868]: DEBUG nova.network.neutron [req-57f309d2-f9d8-4439-a8f4-fe0501fec8f0 req-b82693df-eb27-4c9c-bc49-73fd95b55a56 service nova] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.568326] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.732297] env[61868]: DEBUG nova.network.neutron [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 729.852901] env[61868]: DEBUG nova.network.neutron [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.955293] env[61868]: DEBUG oslo_concurrency.lockutils [req-57f309d2-f9d8-4439-a8f4-fe0501fec8f0 req-b82693df-eb27-4c9c-bc49-73fd95b55a56 service nova] Releasing lock "refresh_cache-66c5f446-91e5-4de1-abb6-f06a0a36020d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.355413] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Releasing lock "refresh_cache-96694bb9-fbf2-4b71-9cb5-37f81b634992" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.355805] env[61868]: DEBUG nova.compute.manager [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 730.355805] env[61868]: DEBUG nova.compute.manager [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 730.355927] env[61868]: DEBUG nova.network.neutron [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 730.376043] env[61868]: DEBUG nova.network.neutron [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 730.609871] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5d7bfc8-74b9-4d8f-9b71-2100e3b9b128 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.618208] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43928b31-ff0b-48fc-919b-2fd211c92147 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.654745] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e90fb0a7-9a65-427a-b3cd-4340bace5916 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.662383] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd68f0d2-48b9-437e-b522-b0ab6cc8450a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.676986] env[61868]: DEBUG nova.compute.provider_tree [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 730.842348] env[61868]: DEBUG nova.compute.manager [req-1778ef6d-6c0d-4fc8-b7d0-5a6eb6a59bb6 req-a6516b00-d7b1-4df5-857e-85ac2a75b4fd service nova] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Received event network-vif-deleted-4c4da9ff-f957-4820-8c63-da666cca0fa6 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 730.879833] env[61868]: DEBUG nova.network.neutron [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.060317] env[61868]: DEBUG nova.compute.manager [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 731.060877] env[61868]: DEBUG nova.virt.hardware [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 731.061121] env[61868]: DEBUG nova.virt.hardware [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 731.061250] env[61868]: DEBUG nova.virt.hardware [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 731.061471] env[61868]: DEBUG nova.virt.hardware [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 731.061861] env[61868]: DEBUG nova.virt.hardware [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 731.061861] env[61868]: DEBUG nova.virt.hardware [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 731.061982] env[61868]: DEBUG nova.virt.hardware [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 731.062162] env[61868]: DEBUG nova.virt.hardware [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 731.062520] env[61868]: DEBUG nova.virt.hardware [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 731.062520] env[61868]: DEBUG nova.virt.hardware [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 731.062658] env[61868]: DEBUG nova.virt.hardware [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 731.063517] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fce7ca32-f926-43ae-a9e7-3a27a630e818 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.071748] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-002829fc-d6bf-4678-bd00-07680be155e5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.085537] env[61868]: ERROR nova.compute.manager [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4c4da9ff-f957-4820-8c63-da666cca0fa6, please check neutron logs for more information. [ 731.085537] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Traceback (most recent call last): [ 731.085537] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 731.085537] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] yield resources [ 731.085537] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 731.085537] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] self.driver.spawn(context, instance, image_meta, [ 731.085537] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 731.085537] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 731.085537] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 731.085537] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] vm_ref = self.build_virtual_machine(instance, [ 731.085537] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 731.085942] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] vif_infos = vmwarevif.get_vif_info(self._session, [ 731.085942] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 731.085942] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] for vif in network_info: [ 731.085942] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 731.085942] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] return self._sync_wrapper(fn, *args, **kwargs) [ 731.085942] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 731.085942] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] self.wait() [ 731.085942] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 731.085942] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] self[:] = self._gt.wait() [ 731.085942] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 731.085942] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] return self._exit_event.wait() [ 731.085942] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 731.085942] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] current.throw(*self._exc) [ 731.086332] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 731.086332] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] result = function(*args, **kwargs) [ 731.086332] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 731.086332] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] return func(*args, **kwargs) [ 731.086332] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 731.086332] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] raise e [ 731.086332] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 731.086332] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] nwinfo = self.network_api.allocate_for_instance( [ 731.086332] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 731.086332] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] created_port_ids = self._update_ports_for_instance( [ 731.086332] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 731.086332] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] with excutils.save_and_reraise_exception(): [ 731.086332] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 731.086688] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] self.force_reraise() [ 731.086688] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 731.086688] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] raise self.value [ 731.086688] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 731.086688] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] updated_port = self._update_port( [ 731.086688] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 731.086688] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] _ensure_no_port_binding_failure(port) [ 731.086688] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 731.086688] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] raise exception.PortBindingFailed(port_id=port['id']) [ 731.086688] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] nova.exception.PortBindingFailed: Binding failed for port 4c4da9ff-f957-4820-8c63-da666cca0fa6, please check neutron logs for more information. [ 731.086688] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] [ 731.086688] env[61868]: INFO nova.compute.manager [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Terminating instance [ 731.087877] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] Acquiring lock "refresh_cache-66c5f446-91e5-4de1-abb6-f06a0a36020d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 731.088046] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] Acquired lock "refresh_cache-66c5f446-91e5-4de1-abb6-f06a0a36020d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.088210] env[61868]: DEBUG nova.network.neutron [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 731.180773] env[61868]: DEBUG nova.scheduler.client.report [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 731.381549] env[61868]: INFO nova.compute.manager [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 96694bb9-fbf2-4b71-9cb5-37f81b634992] Took 1.03 seconds to deallocate network for instance. [ 731.606467] env[61868]: DEBUG nova.network.neutron [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 731.681771] env[61868]: DEBUG nova.network.neutron [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.685944] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.491s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 731.686451] env[61868]: DEBUG nova.compute.manager [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 731.692022] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.415s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.183886] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] Releasing lock "refresh_cache-66c5f446-91e5-4de1-abb6-f06a0a36020d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.184562] env[61868]: DEBUG nova.compute.manager [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 732.184882] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fd903bc8-123d-47a3-b062-18264d0a81f0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.194143] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b84d1425-abbb-4fe9-bdbd-799ccf4420f1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.205358] env[61868]: DEBUG nova.compute.utils [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 732.209670] env[61868]: DEBUG nova.compute.manager [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 732.209798] env[61868]: DEBUG nova.network.neutron [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 732.223551] env[61868]: WARNING nova.virt.vmwareapi.driver [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 66c5f446-91e5-4de1-abb6-f06a0a36020d could not be found. [ 732.223774] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 732.224653] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0c987e47-c34e-4187-9fbf-07121dca5116 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.232865] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73fee32c-0a6e-4286-aee0-2ed52b72eef8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.258421] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 66c5f446-91e5-4de1-abb6-f06a0a36020d could not be found. [ 732.258680] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 732.258859] env[61868]: INFO nova.compute.manager [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Took 0.07 seconds to destroy the instance on the hypervisor. [ 732.259121] env[61868]: DEBUG oslo.service.loopingcall [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 732.259338] env[61868]: DEBUG nova.compute.manager [-] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 732.259433] env[61868]: DEBUG nova.network.neutron [-] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 732.264540] env[61868]: DEBUG nova.policy [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '904fe3b4de514befab69c49e592a650a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '30d321afafe54c07a57476867e4882c8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 732.279692] env[61868]: DEBUG nova.network.neutron [-] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 732.418214] env[61868]: INFO nova.scheduler.client.report [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Deleted allocations for instance 96694bb9-fbf2-4b71-9cb5-37f81b634992 [ 732.592939] env[61868]: DEBUG nova.network.neutron [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Successfully created port: 4d03d50a-044d-4665-9949-cb040fc108ae {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 732.612102] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a02579e8-e466-41d2-84c2-9e8ff3de7d2f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.619526] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf06a4d6-5f47-481b-baa8-3f7011ab4ce0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.650198] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aafe72ac-df4d-4b67-8fde-2f9fb228b650 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.658070] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0faa5714-8fa0-4972-a70b-15db9966f913 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.671270] env[61868]: DEBUG nova.compute.provider_tree [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 732.710850] env[61868]: DEBUG nova.compute.manager [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 732.782409] env[61868]: DEBUG nova.network.neutron [-] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.927534] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b463690d-9eee-4aa1-acfd-16bb65f395bb tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Lock "96694bb9-fbf2-4b71-9cb5-37f81b634992" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 147.453s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.178103] env[61868]: DEBUG nova.scheduler.client.report [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 733.285360] env[61868]: INFO nova.compute.manager [-] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Took 1.03 seconds to deallocate network for instance. [ 733.430562] env[61868]: DEBUG nova.compute.manager [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 733.537395] env[61868]: DEBUG nova.compute.manager [req-5f8c024b-aa56-46ca-a0db-a881519c4e24 req-9c67052b-c0e7-425b-9a81-a917b269ac43 service nova] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Received event network-changed-4d03d50a-044d-4665-9949-cb040fc108ae {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 733.537395] env[61868]: DEBUG nova.compute.manager [req-5f8c024b-aa56-46ca-a0db-a881519c4e24 req-9c67052b-c0e7-425b-9a81-a917b269ac43 service nova] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Refreshing instance network info cache due to event network-changed-4d03d50a-044d-4665-9949-cb040fc108ae. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 733.537395] env[61868]: DEBUG oslo_concurrency.lockutils [req-5f8c024b-aa56-46ca-a0db-a881519c4e24 req-9c67052b-c0e7-425b-9a81-a917b269ac43 service nova] Acquiring lock "refresh_cache-97d1d9c9-214f-4573-9efa-3254ea3f2ee0" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.537395] env[61868]: DEBUG oslo_concurrency.lockutils [req-5f8c024b-aa56-46ca-a0db-a881519c4e24 req-9c67052b-c0e7-425b-9a81-a917b269ac43 service nova] Acquired lock "refresh_cache-97d1d9c9-214f-4573-9efa-3254ea3f2ee0" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.537395] env[61868]: DEBUG nova.network.neutron [req-5f8c024b-aa56-46ca-a0db-a881519c4e24 req-9c67052b-c0e7-425b-9a81-a917b269ac43 service nova] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Refreshing network info cache for port 4d03d50a-044d-4665-9949-cb040fc108ae {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 733.540208] env[61868]: ERROR nova.compute.manager [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4d03d50a-044d-4665-9949-cb040fc108ae, please check neutron logs for more information. [ 733.540208] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 733.540208] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 733.540208] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 733.540208] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 733.540208] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 733.540208] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 733.540208] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 733.540208] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.540208] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 733.540208] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.540208] env[61868]: ERROR nova.compute.manager raise self.value [ 733.540208] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 733.540208] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 733.540208] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 733.540208] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 733.540838] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 733.540838] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 733.540838] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4d03d50a-044d-4665-9949-cb040fc108ae, please check neutron logs for more information. [ 733.540838] env[61868]: ERROR nova.compute.manager [ 733.540838] env[61868]: Traceback (most recent call last): [ 733.540838] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 733.540838] env[61868]: listener.cb(fileno) [ 733.540838] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 733.540838] env[61868]: result = function(*args, **kwargs) [ 733.540838] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 733.540838] env[61868]: return func(*args, **kwargs) [ 733.540838] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 733.540838] env[61868]: raise e [ 733.540838] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 733.540838] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 733.540838] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 733.540838] env[61868]: created_port_ids = self._update_ports_for_instance( [ 733.540838] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 733.540838] env[61868]: with excutils.save_and_reraise_exception(): [ 733.540838] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.540838] env[61868]: self.force_reraise() [ 733.540838] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.540838] env[61868]: raise self.value [ 733.540838] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 733.540838] env[61868]: updated_port = self._update_port( [ 733.540838] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 733.540838] env[61868]: _ensure_no_port_binding_failure(port) [ 733.540838] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 733.540838] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 733.541659] env[61868]: nova.exception.PortBindingFailed: Binding failed for port 4d03d50a-044d-4665-9949-cb040fc108ae, please check neutron logs for more information. [ 733.541659] env[61868]: Removing descriptor: 15 [ 733.682745] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.994s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.683477] env[61868]: ERROR nova.compute.manager [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 80ca8e4e-c979-4622-898c-7901cb81d1dc, please check neutron logs for more information. [ 733.683477] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Traceback (most recent call last): [ 733.683477] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 733.683477] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] self.driver.spawn(context, instance, image_meta, [ 733.683477] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 733.683477] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] self._vmops.spawn(context, instance, image_meta, injected_files, [ 733.683477] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 733.683477] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] vm_ref = self.build_virtual_machine(instance, [ 733.683477] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 733.683477] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] vif_infos = vmwarevif.get_vif_info(self._session, [ 733.683477] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 733.683803] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] for vif in network_info: [ 733.683803] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 733.683803] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] return self._sync_wrapper(fn, *args, **kwargs) [ 733.683803] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 733.683803] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] self.wait() [ 733.683803] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 733.683803] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] self[:] = self._gt.wait() [ 733.683803] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 733.683803] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] return self._exit_event.wait() [ 733.683803] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 733.683803] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] current.throw(*self._exc) [ 733.683803] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 733.683803] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] result = function(*args, **kwargs) [ 733.684145] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 733.684145] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] return func(*args, **kwargs) [ 733.684145] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 733.684145] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] raise e [ 733.684145] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 733.684145] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] nwinfo = self.network_api.allocate_for_instance( [ 733.684145] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 733.684145] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] created_port_ids = self._update_ports_for_instance( [ 733.684145] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 733.684145] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] with excutils.save_and_reraise_exception(): [ 733.684145] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.684145] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] self.force_reraise() [ 733.684145] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.684488] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] raise self.value [ 733.684488] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 733.684488] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] updated_port = self._update_port( [ 733.684488] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 733.684488] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] _ensure_no_port_binding_failure(port) [ 733.684488] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 733.684488] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] raise exception.PortBindingFailed(port_id=port['id']) [ 733.684488] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] nova.exception.PortBindingFailed: Binding failed for port 80ca8e4e-c979-4622-898c-7901cb81d1dc, please check neutron logs for more information. [ 733.684488] env[61868]: ERROR nova.compute.manager [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] [ 733.684488] env[61868]: DEBUG nova.compute.utils [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Binding failed for port 80ca8e4e-c979-4622-898c-7901cb81d1dc, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 733.689453] env[61868]: DEBUG nova.compute.manager [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Build of instance cf7b7da2-380e-4a97-b142-0dd9e5ad3c79 was re-scheduled: Binding failed for port 80ca8e4e-c979-4622-898c-7901cb81d1dc, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 733.689967] env[61868]: DEBUG nova.compute.manager [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 733.690219] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Acquiring lock "refresh_cache-cf7b7da2-380e-4a97-b142-0dd9e5ad3c79" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.692645] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Acquired lock "refresh_cache-cf7b7da2-380e-4a97-b142-0dd9e5ad3c79" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.692645] env[61868]: DEBUG nova.network.neutron [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 733.692645] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.897s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.693061] env[61868]: INFO nova.compute.claims [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 733.722255] env[61868]: DEBUG nova.compute.manager [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 733.755778] env[61868]: DEBUG nova.virt.hardware [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 733.756329] env[61868]: DEBUG nova.virt.hardware [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 733.756329] env[61868]: DEBUG nova.virt.hardware [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 733.756479] env[61868]: DEBUG nova.virt.hardware [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 733.756526] env[61868]: DEBUG nova.virt.hardware [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 733.756662] env[61868]: DEBUG nova.virt.hardware [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 733.756981] env[61868]: DEBUG nova.virt.hardware [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 733.757051] env[61868]: DEBUG nova.virt.hardware [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 733.757178] env[61868]: DEBUG nova.virt.hardware [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 733.757549] env[61868]: DEBUG nova.virt.hardware [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 733.757549] env[61868]: DEBUG nova.virt.hardware [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 733.761015] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57fdc7e8-fedf-41c3-9612-26592de8d05e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.768788] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fbdd6d9-e162-4ff5-9fda-3b84950902d5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.783800] env[61868]: ERROR nova.compute.manager [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4d03d50a-044d-4665-9949-cb040fc108ae, please check neutron logs for more information. [ 733.783800] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Traceback (most recent call last): [ 733.783800] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 733.783800] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] yield resources [ 733.783800] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 733.783800] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] self.driver.spawn(context, instance, image_meta, [ 733.783800] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 733.783800] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 733.783800] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 733.783800] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] vm_ref = self.build_virtual_machine(instance, [ 733.783800] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 733.784218] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] vif_infos = vmwarevif.get_vif_info(self._session, [ 733.784218] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 733.784218] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] for vif in network_info: [ 733.784218] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 733.784218] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] return self._sync_wrapper(fn, *args, **kwargs) [ 733.784218] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 733.784218] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] self.wait() [ 733.784218] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 733.784218] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] self[:] = self._gt.wait() [ 733.784218] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 733.784218] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] return self._exit_event.wait() [ 733.784218] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 733.784218] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] current.throw(*self._exc) [ 733.784605] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 733.784605] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] result = function(*args, **kwargs) [ 733.784605] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 733.784605] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] return func(*args, **kwargs) [ 733.784605] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 733.784605] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] raise e [ 733.784605] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 733.784605] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] nwinfo = self.network_api.allocate_for_instance( [ 733.784605] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 733.784605] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] created_port_ids = self._update_ports_for_instance( [ 733.784605] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 733.784605] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] with excutils.save_and_reraise_exception(): [ 733.784605] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.784965] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] self.force_reraise() [ 733.784965] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.784965] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] raise self.value [ 733.784965] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 733.784965] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] updated_port = self._update_port( [ 733.784965] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 733.784965] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] _ensure_no_port_binding_failure(port) [ 733.784965] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 733.784965] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] raise exception.PortBindingFailed(port_id=port['id']) [ 733.784965] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] nova.exception.PortBindingFailed: Binding failed for port 4d03d50a-044d-4665-9949-cb040fc108ae, please check neutron logs for more information. [ 733.784965] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] [ 733.784965] env[61868]: INFO nova.compute.manager [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Terminating instance [ 733.786139] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] Acquiring lock "refresh_cache-97d1d9c9-214f-4573-9efa-3254ea3f2ee0" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.857654] env[61868]: INFO nova.compute.manager [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Took 0.57 seconds to detach 1 volumes for instance. [ 733.859606] env[61868]: DEBUG nova.compute.claims [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 733.860634] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.955457] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.059683] env[61868]: DEBUG nova.network.neutron [req-5f8c024b-aa56-46ca-a0db-a881519c4e24 req-9c67052b-c0e7-425b-9a81-a917b269ac43 service nova] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 734.139283] env[61868]: DEBUG nova.network.neutron [req-5f8c024b-aa56-46ca-a0db-a881519c4e24 req-9c67052b-c0e7-425b-9a81-a917b269ac43 service nova] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.217657] env[61868]: DEBUG nova.network.neutron [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 734.303120] env[61868]: DEBUG nova.network.neutron [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.642840] env[61868]: DEBUG oslo_concurrency.lockutils [req-5f8c024b-aa56-46ca-a0db-a881519c4e24 req-9c67052b-c0e7-425b-9a81-a917b269ac43 service nova] Releasing lock "refresh_cache-97d1d9c9-214f-4573-9efa-3254ea3f2ee0" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 734.645768] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] Acquired lock "refresh_cache-97d1d9c9-214f-4573-9efa-3254ea3f2ee0" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.646013] env[61868]: DEBUG nova.network.neutron [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 734.807400] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Releasing lock "refresh_cache-cf7b7da2-380e-4a97-b142-0dd9e5ad3c79" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 734.807400] env[61868]: DEBUG nova.compute.manager [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 734.807400] env[61868]: DEBUG nova.compute.manager [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 734.807400] env[61868]: DEBUG nova.network.neutron [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 734.824350] env[61868]: DEBUG nova.network.neutron [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.092331] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquiring lock "0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.092552] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Lock "0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.111078] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Acquiring lock "3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.111322] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Lock "3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.127895] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquiring lock "7788f9f8-c08e-46f8-b204-070215223436" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.128239] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Lock "7788f9f8-c08e-46f8-b204-070215223436" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.140274] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Acquiring lock "52a5afc6-4e79-436c-bc94-b61ca9fb860c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.140556] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Lock "52a5afc6-4e79-436c-bc94-b61ca9fb860c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.155529] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bcd2c46-25c3-49c4-bca5-2d2fbdcc56f0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.164641] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c202f36-caee-4061-82f6-46acb8121b42 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.171026] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Acquiring lock "c2693a43-4ea2-4ab0-8915-2fa544780e3c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.171026] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Lock "c2693a43-4ea2-4ab0-8915-2fa544780e3c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.171940] env[61868]: DEBUG nova.network.neutron [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.199683] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76ca5963-233a-4b3a-a376-76881f1a504f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.207692] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76e90af6-7f3a-4d12-9475-d66a82bc1c27 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.223479] env[61868]: DEBUG nova.compute.provider_tree [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 735.290895] env[61868]: DEBUG nova.network.neutron [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.330098] env[61868]: DEBUG nova.network.neutron [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.565798] env[61868]: DEBUG nova.compute.manager [req-935bebf9-4f23-4798-aaa9-2e7f6448f848 req-2056145a-ea19-4d41-a258-a48d91368630 service nova] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Received event network-vif-deleted-4d03d50a-044d-4665-9949-cb040fc108ae {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 735.727246] env[61868]: DEBUG nova.scheduler.client.report [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 735.794775] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] Releasing lock "refresh_cache-97d1d9c9-214f-4573-9efa-3254ea3f2ee0" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.795235] env[61868]: DEBUG nova.compute.manager [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 735.795427] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 735.795715] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-88e0b0c7-89fb-46bd-9a29-c5d8681cd81a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.804281] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92ffd66d-40ee-4dfc-82d0-31fbcede64d9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.827042] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 97d1d9c9-214f-4573-9efa-3254ea3f2ee0 could not be found. [ 735.827042] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 735.827042] env[61868]: INFO nova.compute.manager [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Took 0.03 seconds to destroy the instance on the hypervisor. [ 735.827161] env[61868]: DEBUG oslo.service.loopingcall [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 735.827334] env[61868]: DEBUG nova.compute.manager [-] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 735.827603] env[61868]: DEBUG nova.network.neutron [-] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 735.832926] env[61868]: INFO nova.compute.manager [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: cf7b7da2-380e-4a97-b142-0dd9e5ad3c79] Took 1.03 seconds to deallocate network for instance. [ 735.849440] env[61868]: DEBUG nova.network.neutron [-] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 736.233055] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.541s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.233587] env[61868]: DEBUG nova.compute.manager [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 736.236013] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.752s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.352340] env[61868]: DEBUG nova.network.neutron [-] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.741175] env[61868]: DEBUG nova.compute.utils [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 736.742522] env[61868]: DEBUG nova.compute.manager [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 736.742522] env[61868]: DEBUG nova.network.neutron [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 736.789379] env[61868]: DEBUG nova.policy [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cce2af4737904177a83d12587d265dc9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1a5f4d9f9ef3491286397b85ff547f36', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 736.854480] env[61868]: INFO nova.compute.manager [-] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Took 1.03 seconds to deallocate network for instance. [ 736.858826] env[61868]: DEBUG nova.compute.claims [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 736.859372] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.863158] env[61868]: INFO nova.scheduler.client.report [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Deleted allocations for instance cf7b7da2-380e-4a97-b142-0dd9e5ad3c79 [ 737.059117] env[61868]: DEBUG nova.network.neutron [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Successfully created port: df04fac4-ceb3-44ad-94b4-c043fbd3868c {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 737.158357] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6e1fb1e-32c9-4a56-839b-cbf1c4e287cb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.166268] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e40c3510-9881-4022-a978-f5c806727054 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.200501] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82cb4467-1a23-4714-b773-82f59c949f8e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.214084] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03131f20-2443-4b52-9f4a-2da5d6ee69b9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.225570] env[61868]: DEBUG nova.compute.provider_tree [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 737.248223] env[61868]: DEBUG nova.compute.manager [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 737.375081] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bfbcf7b1-145a-4d66-959a-85eb70643970 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Lock "cf7b7da2-380e-4a97-b142-0dd9e5ad3c79" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 150.362s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.728632] env[61868]: DEBUG nova.scheduler.client.report [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 737.737252] env[61868]: DEBUG nova.compute.manager [req-ab225a0f-6f07-4b3d-8c22-66b93ab18b17 req-36804ad1-ea6a-4294-a9e4-6581545d01c1 service nova] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Received event network-changed-df04fac4-ceb3-44ad-94b4-c043fbd3868c {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 737.737252] env[61868]: DEBUG nova.compute.manager [req-ab225a0f-6f07-4b3d-8c22-66b93ab18b17 req-36804ad1-ea6a-4294-a9e4-6581545d01c1 service nova] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Refreshing instance network info cache due to event network-changed-df04fac4-ceb3-44ad-94b4-c043fbd3868c. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 737.737252] env[61868]: DEBUG oslo_concurrency.lockutils [req-ab225a0f-6f07-4b3d-8c22-66b93ab18b17 req-36804ad1-ea6a-4294-a9e4-6581545d01c1 service nova] Acquiring lock "refresh_cache-84dc8092-c24a-4587-bfcb-865b6eb129aa" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.737252] env[61868]: DEBUG oslo_concurrency.lockutils [req-ab225a0f-6f07-4b3d-8c22-66b93ab18b17 req-36804ad1-ea6a-4294-a9e4-6581545d01c1 service nova] Acquired lock "refresh_cache-84dc8092-c24a-4587-bfcb-865b6eb129aa" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.737252] env[61868]: DEBUG nova.network.neutron [req-ab225a0f-6f07-4b3d-8c22-66b93ab18b17 req-36804ad1-ea6a-4294-a9e4-6581545d01c1 service nova] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Refreshing network info cache for port df04fac4-ceb3-44ad-94b4-c043fbd3868c {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 737.877805] env[61868]: DEBUG nova.compute.manager [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] [instance: df927c0c-b742-4451-9439-60ea63287b99] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 737.897963] env[61868]: ERROR nova.compute.manager [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port df04fac4-ceb3-44ad-94b4-c043fbd3868c, please check neutron logs for more information. [ 737.897963] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 737.897963] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 737.897963] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 737.897963] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 737.897963] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 737.897963] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 737.897963] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 737.897963] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 737.897963] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 737.897963] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 737.897963] env[61868]: ERROR nova.compute.manager raise self.value [ 737.897963] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 737.897963] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 737.897963] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 737.897963] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 737.898520] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 737.898520] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 737.898520] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port df04fac4-ceb3-44ad-94b4-c043fbd3868c, please check neutron logs for more information. [ 737.898520] env[61868]: ERROR nova.compute.manager [ 737.898520] env[61868]: Traceback (most recent call last): [ 737.898520] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 737.898520] env[61868]: listener.cb(fileno) [ 737.898520] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 737.898520] env[61868]: result = function(*args, **kwargs) [ 737.898520] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 737.898520] env[61868]: return func(*args, **kwargs) [ 737.898520] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 737.898520] env[61868]: raise e [ 737.898520] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 737.898520] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 737.898520] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 737.898520] env[61868]: created_port_ids = self._update_ports_for_instance( [ 737.898520] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 737.898520] env[61868]: with excutils.save_and_reraise_exception(): [ 737.898520] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 737.898520] env[61868]: self.force_reraise() [ 737.898520] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 737.898520] env[61868]: raise self.value [ 737.898520] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 737.898520] env[61868]: updated_port = self._update_port( [ 737.898520] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 737.898520] env[61868]: _ensure_no_port_binding_failure(port) [ 737.898520] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 737.898520] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 737.899458] env[61868]: nova.exception.PortBindingFailed: Binding failed for port df04fac4-ceb3-44ad-94b4-c043fbd3868c, please check neutron logs for more information. [ 737.899458] env[61868]: Removing descriptor: 15 [ 738.234064] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.998s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.234832] env[61868]: ERROR nova.compute.manager [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6f2211c4-2123-49a9-b287-3b10300c560d, please check neutron logs for more information. [ 738.234832] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] Traceback (most recent call last): [ 738.234832] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 738.234832] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] self.driver.spawn(context, instance, image_meta, [ 738.234832] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 738.234832] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] self._vmops.spawn(context, instance, image_meta, injected_files, [ 738.234832] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 738.234832] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] vm_ref = self.build_virtual_machine(instance, [ 738.234832] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 738.234832] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] vif_infos = vmwarevif.get_vif_info(self._session, [ 738.234832] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 738.235212] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] for vif in network_info: [ 738.235212] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 738.235212] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] return self._sync_wrapper(fn, *args, **kwargs) [ 738.235212] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 738.235212] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] self.wait() [ 738.235212] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 738.235212] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] self[:] = self._gt.wait() [ 738.235212] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 738.235212] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] return self._exit_event.wait() [ 738.235212] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 738.235212] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] current.throw(*self._exc) [ 738.235212] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 738.235212] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] result = function(*args, **kwargs) [ 738.235586] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 738.235586] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] return func(*args, **kwargs) [ 738.235586] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 738.235586] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] raise e [ 738.235586] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 738.235586] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] nwinfo = self.network_api.allocate_for_instance( [ 738.235586] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 738.235586] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] created_port_ids = self._update_ports_for_instance( [ 738.235586] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 738.235586] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] with excutils.save_and_reraise_exception(): [ 738.235586] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.235586] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] self.force_reraise() [ 738.235586] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.235951] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] raise self.value [ 738.235951] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 738.235951] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] updated_port = self._update_port( [ 738.235951] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.235951] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] _ensure_no_port_binding_failure(port) [ 738.235951] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.235951] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] raise exception.PortBindingFailed(port_id=port['id']) [ 738.235951] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] nova.exception.PortBindingFailed: Binding failed for port 6f2211c4-2123-49a9-b287-3b10300c560d, please check neutron logs for more information. [ 738.235951] env[61868]: ERROR nova.compute.manager [instance: 460255fb-6589-481b-9770-0a2b548c6194] [ 738.236286] env[61868]: DEBUG nova.compute.utils [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Binding failed for port 6f2211c4-2123-49a9-b287-3b10300c560d, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 738.238191] env[61868]: DEBUG nova.compute.manager [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Build of instance 460255fb-6589-481b-9770-0a2b548c6194 was re-scheduled: Binding failed for port 6f2211c4-2123-49a9-b287-3b10300c560d, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 738.238712] env[61868]: DEBUG nova.compute.manager [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 738.239489] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] Acquiring lock "refresh_cache-460255fb-6589-481b-9770-0a2b548c6194" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.239489] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] Acquired lock "refresh_cache-460255fb-6589-481b-9770-0a2b548c6194" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.239489] env[61868]: DEBUG nova.network.neutron [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 738.240472] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.005s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.256026] env[61868]: DEBUG nova.compute.manager [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 738.264869] env[61868]: DEBUG nova.network.neutron [req-ab225a0f-6f07-4b3d-8c22-66b93ab18b17 req-36804ad1-ea6a-4294-a9e4-6581545d01c1 service nova] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 738.288102] env[61868]: DEBUG nova.virt.hardware [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 738.288102] env[61868]: DEBUG nova.virt.hardware [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 738.288102] env[61868]: DEBUG nova.virt.hardware [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 738.288310] env[61868]: DEBUG nova.virt.hardware [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 738.288310] env[61868]: DEBUG nova.virt.hardware [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 738.288310] env[61868]: DEBUG nova.virt.hardware [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 738.288310] env[61868]: DEBUG nova.virt.hardware [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 738.288435] env[61868]: DEBUG nova.virt.hardware [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 738.288963] env[61868]: DEBUG nova.virt.hardware [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 738.288963] env[61868]: DEBUG nova.virt.hardware [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 738.289070] env[61868]: DEBUG nova.virt.hardware [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 738.291079] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac38ef9d-25b7-4f69-b2fa-6f829d449403 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.299618] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab0b193d-3fa7-4332-83dc-346d64a21c5d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.314376] env[61868]: ERROR nova.compute.manager [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port df04fac4-ceb3-44ad-94b4-c043fbd3868c, please check neutron logs for more information. [ 738.314376] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Traceback (most recent call last): [ 738.314376] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 738.314376] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] yield resources [ 738.314376] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 738.314376] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] self.driver.spawn(context, instance, image_meta, [ 738.314376] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 738.314376] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] self._vmops.spawn(context, instance, image_meta, injected_files, [ 738.314376] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 738.314376] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] vm_ref = self.build_virtual_machine(instance, [ 738.314376] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 738.314808] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] vif_infos = vmwarevif.get_vif_info(self._session, [ 738.314808] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 738.314808] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] for vif in network_info: [ 738.314808] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 738.314808] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] return self._sync_wrapper(fn, *args, **kwargs) [ 738.314808] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 738.314808] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] self.wait() [ 738.314808] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 738.314808] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] self[:] = self._gt.wait() [ 738.314808] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 738.314808] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] return self._exit_event.wait() [ 738.314808] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 738.314808] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] current.throw(*self._exc) [ 738.315291] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 738.315291] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] result = function(*args, **kwargs) [ 738.315291] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 738.315291] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] return func(*args, **kwargs) [ 738.315291] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 738.315291] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] raise e [ 738.315291] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 738.315291] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] nwinfo = self.network_api.allocate_for_instance( [ 738.315291] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 738.315291] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] created_port_ids = self._update_ports_for_instance( [ 738.315291] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 738.315291] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] with excutils.save_and_reraise_exception(): [ 738.315291] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.315720] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] self.force_reraise() [ 738.315720] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.315720] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] raise self.value [ 738.315720] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 738.315720] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] updated_port = self._update_port( [ 738.315720] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.315720] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] _ensure_no_port_binding_failure(port) [ 738.315720] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.315720] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] raise exception.PortBindingFailed(port_id=port['id']) [ 738.315720] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] nova.exception.PortBindingFailed: Binding failed for port df04fac4-ceb3-44ad-94b4-c043fbd3868c, please check neutron logs for more information. [ 738.315720] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] [ 738.315720] env[61868]: INFO nova.compute.manager [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Terminating instance [ 738.316650] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Acquiring lock "refresh_cache-84dc8092-c24a-4587-bfcb-865b6eb129aa" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.379992] env[61868]: DEBUG nova.network.neutron [req-ab225a0f-6f07-4b3d-8c22-66b93ab18b17 req-36804ad1-ea6a-4294-a9e4-6581545d01c1 service nova] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.401447] env[61868]: DEBUG oslo_concurrency.lockutils [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.773245] env[61868]: DEBUG nova.network.neutron [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 738.888429] env[61868]: DEBUG oslo_concurrency.lockutils [req-ab225a0f-6f07-4b3d-8c22-66b93ab18b17 req-36804ad1-ea6a-4294-a9e4-6581545d01c1 service nova] Releasing lock "refresh_cache-84dc8092-c24a-4587-bfcb-865b6eb129aa" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.888429] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Acquired lock "refresh_cache-84dc8092-c24a-4587-bfcb-865b6eb129aa" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.888429] env[61868]: DEBUG nova.network.neutron [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 738.895907] env[61868]: DEBUG nova.network.neutron [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.152118] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72be53f3-cbe3-471d-aa64-8f81b9704071 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.160304] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41ed818a-4950-4655-ab13-446a68b34ff7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.191793] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98defb7a-ad64-483d-ba15-b693213834d7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.198978] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0546fe26-e650-4b9f-a2be-f5bf717c756b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.212181] env[61868]: DEBUG nova.compute.provider_tree [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 739.401803] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] Releasing lock "refresh_cache-460255fb-6589-481b-9770-0a2b548c6194" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.402065] env[61868]: DEBUG nova.compute.manager [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 739.402256] env[61868]: DEBUG nova.compute.manager [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 739.402419] env[61868]: DEBUG nova.network.neutron [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] [instance: 460255fb-6589-481b-9770-0a2b548c6194] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 739.404847] env[61868]: DEBUG nova.network.neutron [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 739.420448] env[61868]: DEBUG nova.network.neutron [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 739.487645] env[61868]: DEBUG nova.network.neutron [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.715962] env[61868]: DEBUG nova.scheduler.client.report [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 739.799885] env[61868]: DEBUG nova.compute.manager [req-d5c8e4e2-97c1-4fe2-809a-445888f71b4c req-b0ca5f96-f53f-49c6-ac41-9469eb4bd7f6 service nova] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Received event network-vif-deleted-df04fac4-ceb3-44ad-94b4-c043fbd3868c {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 739.923416] env[61868]: DEBUG nova.network.neutron [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.990416] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Releasing lock "refresh_cache-84dc8092-c24a-4587-bfcb-865b6eb129aa" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.990878] env[61868]: DEBUG nova.compute.manager [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 739.991095] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 739.991382] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-768cb244-ed68-402a-9769-5a4ba8bcddd1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.000240] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bad7e7a8-f737-4af0-b671-5e8d022dc52c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.021433] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 84dc8092-c24a-4587-bfcb-865b6eb129aa could not be found. [ 740.021644] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 740.021822] env[61868]: INFO nova.compute.manager [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Took 0.03 seconds to destroy the instance on the hypervisor. [ 740.022070] env[61868]: DEBUG oslo.service.loopingcall [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 740.022284] env[61868]: DEBUG nova.compute.manager [-] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 740.022414] env[61868]: DEBUG nova.network.neutron [-] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 740.037062] env[61868]: DEBUG nova.network.neutron [-] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.221202] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.980s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.222102] env[61868]: ERROR nova.compute.manager [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9fe37d4c-b367-48ea-b471-3a851ec07246, please check neutron logs for more information. [ 740.222102] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Traceback (most recent call last): [ 740.222102] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 740.222102] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] self.driver.spawn(context, instance, image_meta, [ 740.222102] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 740.222102] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 740.222102] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 740.222102] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] vm_ref = self.build_virtual_machine(instance, [ 740.222102] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 740.222102] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] vif_infos = vmwarevif.get_vif_info(self._session, [ 740.222102] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 740.222502] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] for vif in network_info: [ 740.222502] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 740.222502] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] return self._sync_wrapper(fn, *args, **kwargs) [ 740.222502] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 740.222502] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] self.wait() [ 740.222502] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 740.222502] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] self[:] = self._gt.wait() [ 740.222502] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 740.222502] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] return self._exit_event.wait() [ 740.222502] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 740.222502] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] current.throw(*self._exc) [ 740.222502] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 740.222502] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] result = function(*args, **kwargs) [ 740.222851] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 740.222851] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] return func(*args, **kwargs) [ 740.222851] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 740.222851] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] raise e [ 740.222851] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 740.222851] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] nwinfo = self.network_api.allocate_for_instance( [ 740.222851] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 740.222851] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] created_port_ids = self._update_ports_for_instance( [ 740.222851] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 740.222851] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] with excutils.save_and_reraise_exception(): [ 740.222851] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 740.222851] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] self.force_reraise() [ 740.222851] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 740.223259] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] raise self.value [ 740.223259] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 740.223259] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] updated_port = self._update_port( [ 740.223259] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 740.223259] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] _ensure_no_port_binding_failure(port) [ 740.223259] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 740.223259] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] raise exception.PortBindingFailed(port_id=port['id']) [ 740.223259] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] nova.exception.PortBindingFailed: Binding failed for port 9fe37d4c-b367-48ea-b471-3a851ec07246, please check neutron logs for more information. [ 740.223259] env[61868]: ERROR nova.compute.manager [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] [ 740.223259] env[61868]: DEBUG nova.compute.utils [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Binding failed for port 9fe37d4c-b367-48ea-b471-3a851ec07246, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 740.224939] env[61868]: DEBUG oslo_concurrency.lockutils [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.288s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.228246] env[61868]: DEBUG nova.compute.manager [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Build of instance 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5 was re-scheduled: Binding failed for port 9fe37d4c-b367-48ea-b471-3a851ec07246, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 740.228878] env[61868]: DEBUG nova.compute.manager [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 740.229092] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "refresh_cache-30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.229205] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquired lock "refresh_cache-30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.229436] env[61868]: DEBUG nova.network.neutron [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 740.248344] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Acquiring lock "409cad1a-946d-4c58-aa57-1c0bf97fe63f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.248344] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Lock "409cad1a-946d-4c58-aa57-1c0bf97fe63f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.427121] env[61868]: INFO nova.compute.manager [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] [instance: 460255fb-6589-481b-9770-0a2b548c6194] Took 1.02 seconds to deallocate network for instance. [ 740.542059] env[61868]: DEBUG nova.network.neutron [-] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.750443] env[61868]: DEBUG nova.network.neutron [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.840903] env[61868]: DEBUG nova.network.neutron [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.045500] env[61868]: INFO nova.compute.manager [-] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Took 1.02 seconds to deallocate network for instance. [ 741.047726] env[61868]: DEBUG nova.compute.claims [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 741.047900] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.062943] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83f62760-6f20-4c7a-89d1-8afa7ea7b899 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.070418] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae7cf5e1-2e05-47eb-974e-ef0f541461e6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.099504] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c748cbc-3a39-45db-8de2-57abd2e11f7e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.106745] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d2469aa-a30b-4a42-83cb-6422102d850a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.119421] env[61868]: DEBUG nova.compute.provider_tree [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 741.344198] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Releasing lock "refresh_cache-30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.344198] env[61868]: DEBUG nova.compute.manager [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 741.344198] env[61868]: DEBUG nova.compute.manager [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 741.344198] env[61868]: DEBUG nova.network.neutron [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 741.361415] env[61868]: DEBUG nova.network.neutron [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 741.454921] env[61868]: INFO nova.scheduler.client.report [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] Deleted allocations for instance 460255fb-6589-481b-9770-0a2b548c6194 [ 741.623022] env[61868]: DEBUG nova.scheduler.client.report [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 741.864588] env[61868]: DEBUG nova.network.neutron [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.964683] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9bf6ae81-eb93-474f-bd66-9ad6735798ee tempest-AttachInterfacesV270Test-1512318083 tempest-AttachInterfacesV270Test-1512318083-project-member] Lock "460255fb-6589-481b-9770-0a2b548c6194" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 153.606s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.128038] env[61868]: DEBUG oslo_concurrency.lockutils [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.903s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.128681] env[61868]: ERROR nova.compute.manager [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5a36ed2f-6282-43a3-840f-9ab052883ad3, please check neutron logs for more information. [ 742.128681] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Traceback (most recent call last): [ 742.128681] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 742.128681] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] self.driver.spawn(context, instance, image_meta, [ 742.128681] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 742.128681] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 742.128681] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 742.128681] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] vm_ref = self.build_virtual_machine(instance, [ 742.128681] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 742.128681] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] vif_infos = vmwarevif.get_vif_info(self._session, [ 742.128681] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 742.129230] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] for vif in network_info: [ 742.129230] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 742.129230] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] return self._sync_wrapper(fn, *args, **kwargs) [ 742.129230] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 742.129230] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] self.wait() [ 742.129230] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 742.129230] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] self[:] = self._gt.wait() [ 742.129230] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 742.129230] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] return self._exit_event.wait() [ 742.129230] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 742.129230] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] current.throw(*self._exc) [ 742.129230] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 742.129230] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] result = function(*args, **kwargs) [ 742.129592] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 742.129592] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] return func(*args, **kwargs) [ 742.129592] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 742.129592] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] raise e [ 742.129592] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 742.129592] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] nwinfo = self.network_api.allocate_for_instance( [ 742.129592] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 742.129592] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] created_port_ids = self._update_ports_for_instance( [ 742.129592] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 742.129592] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] with excutils.save_and_reraise_exception(): [ 742.129592] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 742.129592] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] self.force_reraise() [ 742.129592] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 742.129935] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] raise self.value [ 742.129935] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 742.129935] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] updated_port = self._update_port( [ 742.129935] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 742.129935] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] _ensure_no_port_binding_failure(port) [ 742.129935] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 742.129935] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] raise exception.PortBindingFailed(port_id=port['id']) [ 742.129935] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] nova.exception.PortBindingFailed: Binding failed for port 5a36ed2f-6282-43a3-840f-9ab052883ad3, please check neutron logs for more information. [ 742.129935] env[61868]: ERROR nova.compute.manager [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] [ 742.129935] env[61868]: DEBUG nova.compute.utils [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Binding failed for port 5a36ed2f-6282-43a3-840f-9ab052883ad3, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 742.131955] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.013s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.134023] env[61868]: DEBUG nova.compute.manager [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Build of instance b4a7ff58-0f48-4942-a270-de6910ff0c0c was re-scheduled: Binding failed for port 5a36ed2f-6282-43a3-840f-9ab052883ad3, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 742.134202] env[61868]: DEBUG nova.compute.manager [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 742.134402] env[61868]: DEBUG oslo_concurrency.lockutils [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Acquiring lock "refresh_cache-b4a7ff58-0f48-4942-a270-de6910ff0c0c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.134547] env[61868]: DEBUG oslo_concurrency.lockutils [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Acquired lock "refresh_cache-b4a7ff58-0f48-4942-a270-de6910ff0c0c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.134706] env[61868]: DEBUG nova.network.neutron [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 742.367753] env[61868]: INFO nova.compute.manager [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5] Took 1.02 seconds to deallocate network for instance. [ 742.467278] env[61868]: DEBUG nova.compute.manager [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 742.663171] env[61868]: DEBUG nova.network.neutron [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 742.784455] env[61868]: DEBUG nova.network.neutron [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.992019] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 743.031383] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ead63a8-9b1a-40aa-bdb8-459cb72a0421 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.039238] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c643df0f-7d12-43f4-bb23-ce665c3de068 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.071714] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c437cc1d-33a1-4096-9c4a-81b9cf930ab5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.079342] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0a16932-b6cf-4312-8b88-10e0ddd89406 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.092411] env[61868]: DEBUG nova.compute.provider_tree [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 743.287341] env[61868]: DEBUG oslo_concurrency.lockutils [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Releasing lock "refresh_cache-b4a7ff58-0f48-4942-a270-de6910ff0c0c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.287675] env[61868]: DEBUG nova.compute.manager [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 743.287774] env[61868]: DEBUG nova.compute.manager [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 743.287941] env[61868]: DEBUG nova.network.neutron [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 743.303666] env[61868]: DEBUG nova.network.neutron [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 743.404878] env[61868]: INFO nova.scheduler.client.report [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Deleted allocations for instance 30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5 [ 743.596029] env[61868]: DEBUG nova.scheduler.client.report [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 743.806807] env[61868]: DEBUG nova.network.neutron [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.913219] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da0bb88d-820a-4cf4-9c7e-e2a4ac8e3c6e tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "30cf9ebc-daab-42d3-9e7d-cb4d2cb55bf5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 149.135s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.100746] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.970s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.101424] env[61868]: ERROR nova.compute.manager [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9967a7d8-aea8-47aa-b6e5-2d7eb28a285a, please check neutron logs for more information. [ 744.101424] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Traceback (most recent call last): [ 744.101424] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 744.101424] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] self.driver.spawn(context, instance, image_meta, [ 744.101424] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 744.101424] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 744.101424] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 744.101424] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] vm_ref = self.build_virtual_machine(instance, [ 744.101424] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 744.101424] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] vif_infos = vmwarevif.get_vif_info(self._session, [ 744.101424] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 744.101808] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] for vif in network_info: [ 744.101808] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 744.101808] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] return self._sync_wrapper(fn, *args, **kwargs) [ 744.101808] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 744.101808] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] self.wait() [ 744.101808] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 744.101808] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] self[:] = self._gt.wait() [ 744.101808] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 744.101808] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] return self._exit_event.wait() [ 744.101808] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 744.101808] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] current.throw(*self._exc) [ 744.101808] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 744.101808] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] result = function(*args, **kwargs) [ 744.102230] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 744.102230] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] return func(*args, **kwargs) [ 744.102230] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 744.102230] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] raise e [ 744.102230] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 744.102230] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] nwinfo = self.network_api.allocate_for_instance( [ 744.102230] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 744.102230] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] created_port_ids = self._update_ports_for_instance( [ 744.102230] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 744.102230] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] with excutils.save_and_reraise_exception(): [ 744.102230] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 744.102230] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] self.force_reraise() [ 744.102230] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 744.103336] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] raise self.value [ 744.103336] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 744.103336] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] updated_port = self._update_port( [ 744.103336] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 744.103336] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] _ensure_no_port_binding_failure(port) [ 744.103336] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 744.103336] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] raise exception.PortBindingFailed(port_id=port['id']) [ 744.103336] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] nova.exception.PortBindingFailed: Binding failed for port 9967a7d8-aea8-47aa-b6e5-2d7eb28a285a, please check neutron logs for more information. [ 744.103336] env[61868]: ERROR nova.compute.manager [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] [ 744.103336] env[61868]: DEBUG nova.compute.utils [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Binding failed for port 9967a7d8-aea8-47aa-b6e5-2d7eb28a285a, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 744.103663] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.538s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.104795] env[61868]: INFO nova.compute.claims [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 744.108086] env[61868]: DEBUG nova.compute.manager [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Build of instance 92d3557c-655f-45d9-ae7d-bbe62567e1e2 was re-scheduled: Binding failed for port 9967a7d8-aea8-47aa-b6e5-2d7eb28a285a, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 744.108569] env[61868]: DEBUG nova.compute.manager [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 744.108797] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Acquiring lock "refresh_cache-92d3557c-655f-45d9-ae7d-bbe62567e1e2" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.108943] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Acquired lock "refresh_cache-92d3557c-655f-45d9-ae7d-bbe62567e1e2" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.109124] env[61868]: DEBUG nova.network.neutron [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 744.310026] env[61868]: INFO nova.compute.manager [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] [instance: b4a7ff58-0f48-4942-a270-de6910ff0c0c] Took 1.02 seconds to deallocate network for instance. [ 744.415550] env[61868]: DEBUG nova.compute.manager [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 744.636419] env[61868]: DEBUG nova.network.neutron [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 744.700304] env[61868]: DEBUG nova.network.neutron [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.932672] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 745.039666] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "c7b10943-5136-44c8-b4e7-59651220a333" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 745.040257] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "c7b10943-5136-44c8-b4e7-59651220a333" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.204779] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Releasing lock "refresh_cache-92d3557c-655f-45d9-ae7d-bbe62567e1e2" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.205078] env[61868]: DEBUG nova.compute.manager [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 745.205286] env[61868]: DEBUG nova.compute.manager [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 745.205480] env[61868]: DEBUG nova.network.neutron [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 745.221586] env[61868]: DEBUG nova.network.neutron [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 745.337107] env[61868]: INFO nova.scheduler.client.report [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Deleted allocations for instance b4a7ff58-0f48-4942-a270-de6910ff0c0c [ 745.458422] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-596cf3d5-7072-4987-9415-869ef1499ec7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.466082] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d8eb5e7-f95b-4864-b09d-d83a92670ce3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.495163] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6693e2e-d9cf-4cca-9c4f-5d69167c29ec {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.502290] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93128c37-4986-4f72-a597-5ec7382ecf44 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.516184] env[61868]: DEBUG nova.compute.provider_tree [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 745.723935] env[61868]: DEBUG nova.network.neutron [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.847429] env[61868]: DEBUG oslo_concurrency.lockutils [None req-511c5983-b9b3-4b7a-b7ea-e7b62f0a10ce tempest-ServersTestMultiNic-1616910989 tempest-ServersTestMultiNic-1616910989-project-member] Lock "b4a7ff58-0f48-4942-a270-de6910ff0c0c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 149.308s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.021945] env[61868]: DEBUG nova.scheduler.client.report [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 746.226952] env[61868]: INFO nova.compute.manager [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] [instance: 92d3557c-655f-45d9-ae7d-bbe62567e1e2] Took 1.02 seconds to deallocate network for instance. [ 746.350549] env[61868]: DEBUG nova.compute.manager [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 746.527285] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.424s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.527891] env[61868]: DEBUG nova.compute.manager [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 746.531020] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.671s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.876313] env[61868]: DEBUG oslo_concurrency.lockutils [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.036038] env[61868]: DEBUG nova.compute.utils [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 747.046012] env[61868]: DEBUG nova.compute.manager [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 747.046282] env[61868]: DEBUG nova.network.neutron [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 747.138229] env[61868]: DEBUG nova.policy [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cce2af4737904177a83d12587d265dc9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1a5f4d9f9ef3491286397b85ff547f36', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 747.267842] env[61868]: INFO nova.scheduler.client.report [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Deleted allocations for instance 92d3557c-655f-45d9-ae7d-bbe62567e1e2 [ 747.457379] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-837e15c7-8160-4bae-9239-4a794756829f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.465661] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e357ffc-7df5-4a2b-9b5b-5c7864f1dee1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.497564] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82d6afc0-0327-47ba-9bec-49ab5b6e06a9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.506471] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e47fc5ef-dc22-4989-91fe-b85ed4d9cb07 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.523437] env[61868]: DEBUG nova.compute.provider_tree [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 747.546344] env[61868]: DEBUG nova.compute.manager [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 747.696332] env[61868]: DEBUG nova.network.neutron [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Successfully created port: 898e66dc-6927-44a5-9a0a-1a0114d58d14 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 747.779638] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5d37c1de-58d0-49e1-b9b5-1cfd7bd83780 tempest-MigrationsAdminTest-1814864180 tempest-MigrationsAdminTest-1814864180-project-member] Lock "92d3557c-655f-45d9-ae7d-bbe62567e1e2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 150.866s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.027158] env[61868]: DEBUG nova.scheduler.client.report [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 748.283653] env[61868]: DEBUG nova.compute.manager [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 748.532176] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.533073] env[61868]: ERROR nova.compute.manager [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4c4da9ff-f957-4820-8c63-da666cca0fa6, please check neutron logs for more information. [ 748.533073] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Traceback (most recent call last): [ 748.533073] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 748.533073] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] self.driver.spawn(context, instance, image_meta, [ 748.533073] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 748.533073] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 748.533073] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 748.533073] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] vm_ref = self.build_virtual_machine(instance, [ 748.533073] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 748.533073] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] vif_infos = vmwarevif.get_vif_info(self._session, [ 748.533073] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 748.533482] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] for vif in network_info: [ 748.533482] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 748.533482] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] return self._sync_wrapper(fn, *args, **kwargs) [ 748.533482] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 748.533482] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] self.wait() [ 748.533482] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 748.533482] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] self[:] = self._gt.wait() [ 748.533482] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 748.533482] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] return self._exit_event.wait() [ 748.533482] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 748.533482] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] current.throw(*self._exc) [ 748.533482] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 748.533482] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] result = function(*args, **kwargs) [ 748.533856] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 748.533856] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] return func(*args, **kwargs) [ 748.533856] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 748.533856] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] raise e [ 748.533856] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 748.533856] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] nwinfo = self.network_api.allocate_for_instance( [ 748.533856] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 748.533856] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] created_port_ids = self._update_ports_for_instance( [ 748.533856] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 748.533856] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] with excutils.save_and_reraise_exception(): [ 748.533856] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 748.533856] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] self.force_reraise() [ 748.533856] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 748.534242] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] raise self.value [ 748.534242] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 748.534242] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] updated_port = self._update_port( [ 748.534242] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 748.534242] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] _ensure_no_port_binding_failure(port) [ 748.534242] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 748.534242] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] raise exception.PortBindingFailed(port_id=port['id']) [ 748.534242] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] nova.exception.PortBindingFailed: Binding failed for port 4c4da9ff-f957-4820-8c63-da666cca0fa6, please check neutron logs for more information. [ 748.534242] env[61868]: ERROR nova.compute.manager [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] [ 748.534242] env[61868]: DEBUG nova.compute.utils [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Binding failed for port 4c4da9ff-f957-4820-8c63-da666cca0fa6, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 748.535069] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.580s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.536701] env[61868]: INFO nova.compute.claims [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 748.542175] env[61868]: DEBUG nova.compute.manager [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Build of instance 66c5f446-91e5-4de1-abb6-f06a0a36020d was re-scheduled: Binding failed for port 4c4da9ff-f957-4820-8c63-da666cca0fa6, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 748.542175] env[61868]: DEBUG nova.compute.manager [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 748.542175] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] Acquiring lock "refresh_cache-66c5f446-91e5-4de1-abb6-f06a0a36020d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.542175] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] Acquired lock "refresh_cache-66c5f446-91e5-4de1-abb6-f06a0a36020d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.542396] env[61868]: DEBUG nova.network.neutron [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 748.556861] env[61868]: DEBUG nova.compute.manager [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 748.589033] env[61868]: DEBUG nova.virt.hardware [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 748.589282] env[61868]: DEBUG nova.virt.hardware [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 748.589437] env[61868]: DEBUG nova.virt.hardware [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 748.589615] env[61868]: DEBUG nova.virt.hardware [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 748.589781] env[61868]: DEBUG nova.virt.hardware [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 748.589917] env[61868]: DEBUG nova.virt.hardware [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 748.591925] env[61868]: DEBUG nova.virt.hardware [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 748.591925] env[61868]: DEBUG nova.virt.hardware [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 748.591925] env[61868]: DEBUG nova.virt.hardware [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 748.591925] env[61868]: DEBUG nova.virt.hardware [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 748.591925] env[61868]: DEBUG nova.virt.hardware [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 748.592840] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-732625f6-00d4-464d-9f83-5e1ac9ba22b8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.602763] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fb48c19-7714-4625-a2db-15d319bb33e0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.814648] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.821333] env[61868]: DEBUG nova.compute.manager [req-fc18c32b-4272-4fc9-91a2-cd04ee8e76fd req-8c0eb88e-9b07-4738-9314-859b39ce3e7d service nova] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Received event network-changed-898e66dc-6927-44a5-9a0a-1a0114d58d14 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 748.821333] env[61868]: DEBUG nova.compute.manager [req-fc18c32b-4272-4fc9-91a2-cd04ee8e76fd req-8c0eb88e-9b07-4738-9314-859b39ce3e7d service nova] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Refreshing instance network info cache due to event network-changed-898e66dc-6927-44a5-9a0a-1a0114d58d14. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 748.821333] env[61868]: DEBUG oslo_concurrency.lockutils [req-fc18c32b-4272-4fc9-91a2-cd04ee8e76fd req-8c0eb88e-9b07-4738-9314-859b39ce3e7d service nova] Acquiring lock "refresh_cache-06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.821333] env[61868]: DEBUG oslo_concurrency.lockutils [req-fc18c32b-4272-4fc9-91a2-cd04ee8e76fd req-8c0eb88e-9b07-4738-9314-859b39ce3e7d service nova] Acquired lock "refresh_cache-06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.821333] env[61868]: DEBUG nova.network.neutron [req-fc18c32b-4272-4fc9-91a2-cd04ee8e76fd req-8c0eb88e-9b07-4738-9314-859b39ce3e7d service nova] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Refreshing network info cache for port 898e66dc-6927-44a5-9a0a-1a0114d58d14 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 749.075038] env[61868]: DEBUG nova.network.neutron [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 749.302141] env[61868]: DEBUG nova.network.neutron [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.350238] env[61868]: DEBUG nova.network.neutron [req-fc18c32b-4272-4fc9-91a2-cd04ee8e76fd req-8c0eb88e-9b07-4738-9314-859b39ce3e7d service nova] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 749.400118] env[61868]: ERROR nova.compute.manager [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 898e66dc-6927-44a5-9a0a-1a0114d58d14, please check neutron logs for more information. [ 749.400118] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 749.400118] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 749.400118] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 749.400118] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 749.400118] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 749.400118] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 749.400118] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 749.400118] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.400118] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 749.400118] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.400118] env[61868]: ERROR nova.compute.manager raise self.value [ 749.400118] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 749.400118] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 749.400118] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.400118] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 749.400719] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.400719] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 749.400719] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 898e66dc-6927-44a5-9a0a-1a0114d58d14, please check neutron logs for more information. [ 749.400719] env[61868]: ERROR nova.compute.manager [ 749.400719] env[61868]: Traceback (most recent call last): [ 749.400719] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 749.400719] env[61868]: listener.cb(fileno) [ 749.400719] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 749.400719] env[61868]: result = function(*args, **kwargs) [ 749.400719] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 749.400719] env[61868]: return func(*args, **kwargs) [ 749.400719] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 749.400719] env[61868]: raise e [ 749.400719] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 749.400719] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 749.400719] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 749.400719] env[61868]: created_port_ids = self._update_ports_for_instance( [ 749.400719] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 749.400719] env[61868]: with excutils.save_and_reraise_exception(): [ 749.400719] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.400719] env[61868]: self.force_reraise() [ 749.400719] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.400719] env[61868]: raise self.value [ 749.400719] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 749.400719] env[61868]: updated_port = self._update_port( [ 749.400719] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.400719] env[61868]: _ensure_no_port_binding_failure(port) [ 749.400719] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.400719] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 749.401701] env[61868]: nova.exception.PortBindingFailed: Binding failed for port 898e66dc-6927-44a5-9a0a-1a0114d58d14, please check neutron logs for more information. [ 749.401701] env[61868]: Removing descriptor: 15 [ 749.401701] env[61868]: ERROR nova.compute.manager [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 898e66dc-6927-44a5-9a0a-1a0114d58d14, please check neutron logs for more information. [ 749.401701] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Traceback (most recent call last): [ 749.401701] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 749.401701] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] yield resources [ 749.401701] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 749.401701] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] self.driver.spawn(context, instance, image_meta, [ 749.401701] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 749.401701] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] self._vmops.spawn(context, instance, image_meta, injected_files, [ 749.401701] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 749.401701] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] vm_ref = self.build_virtual_machine(instance, [ 749.402178] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 749.402178] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] vif_infos = vmwarevif.get_vif_info(self._session, [ 749.402178] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 749.402178] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] for vif in network_info: [ 749.402178] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 749.402178] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] return self._sync_wrapper(fn, *args, **kwargs) [ 749.402178] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 749.402178] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] self.wait() [ 749.402178] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 749.402178] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] self[:] = self._gt.wait() [ 749.402178] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 749.402178] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] return self._exit_event.wait() [ 749.402178] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 749.402592] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] result = hub.switch() [ 749.402592] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 749.402592] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] return self.greenlet.switch() [ 749.402592] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 749.402592] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] result = function(*args, **kwargs) [ 749.402592] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 749.402592] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] return func(*args, **kwargs) [ 749.402592] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 749.402592] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] raise e [ 749.402592] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 749.402592] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] nwinfo = self.network_api.allocate_for_instance( [ 749.402592] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 749.402592] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] created_port_ids = self._update_ports_for_instance( [ 749.402992] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 749.402992] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] with excutils.save_and_reraise_exception(): [ 749.402992] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.402992] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] self.force_reraise() [ 749.402992] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.402992] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] raise self.value [ 749.402992] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 749.402992] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] updated_port = self._update_port( [ 749.402992] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.402992] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] _ensure_no_port_binding_failure(port) [ 749.402992] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.402992] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] raise exception.PortBindingFailed(port_id=port['id']) [ 749.403408] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] nova.exception.PortBindingFailed: Binding failed for port 898e66dc-6927-44a5-9a0a-1a0114d58d14, please check neutron logs for more information. [ 749.403408] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] [ 749.403408] env[61868]: INFO nova.compute.manager [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Terminating instance [ 749.403408] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Acquiring lock "refresh_cache-06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.450479] env[61868]: DEBUG nova.network.neutron [req-fc18c32b-4272-4fc9-91a2-cd04ee8e76fd req-8c0eb88e-9b07-4738-9314-859b39ce3e7d service nova] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.804979] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] Releasing lock "refresh_cache-66c5f446-91e5-4de1-abb6-f06a0a36020d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 749.805463] env[61868]: DEBUG nova.compute.manager [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 749.806303] env[61868]: DEBUG nova.compute.manager [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 749.807493] env[61868]: DEBUG nova.network.neutron [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 749.826098] env[61868]: DEBUG nova.network.neutron [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 749.930545] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a991ef8-d78c-495b-84da-0c573a196f88 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.940213] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-718ebb73-1d3f-4db9-9676-3c83e469c576 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.970527] env[61868]: DEBUG oslo_concurrency.lockutils [req-fc18c32b-4272-4fc9-91a2-cd04ee8e76fd req-8c0eb88e-9b07-4738-9314-859b39ce3e7d service nova] Releasing lock "refresh_cache-06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 749.971433] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Acquired lock "refresh_cache-06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.971622] env[61868]: DEBUG nova.network.neutron [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 749.975408] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bee33a1-2f87-4911-a310-26855118c431 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.982640] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2331f6c-3253-45da-b017-51c490b306b7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.999688] env[61868]: DEBUG nova.compute.provider_tree [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 750.328657] env[61868]: DEBUG nova.network.neutron [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.495985] env[61868]: DEBUG nova.network.neutron [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 750.502905] env[61868]: DEBUG nova.scheduler.client.report [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 750.618049] env[61868]: DEBUG nova.network.neutron [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.833169] env[61868]: INFO nova.compute.manager [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] [instance: 66c5f446-91e5-4de1-abb6-f06a0a36020d] Took 1.03 seconds to deallocate network for instance. [ 750.878925] env[61868]: DEBUG nova.compute.manager [req-f25be16d-ca95-452e-adbc-5ee10887226e req-12db7066-a3b9-446f-a2ce-63da2cdf6e4f service nova] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Received event network-vif-deleted-898e66dc-6927-44a5-9a0a-1a0114d58d14 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 751.007703] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.473s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.008271] env[61868]: DEBUG nova.compute.manager [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 751.015772] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.157s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.120830] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Releasing lock "refresh_cache-06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.121339] env[61868]: DEBUG nova.compute.manager [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 751.121534] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 751.121837] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9599e45b-e788-4024-8ffd-bbb634bd963c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.130854] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53705f21-3753-4bd0-a20b-7ca194139694 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.152744] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea could not be found. [ 751.153091] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 751.153300] env[61868]: INFO nova.compute.manager [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Took 0.03 seconds to destroy the instance on the hypervisor. [ 751.153553] env[61868]: DEBUG oslo.service.loopingcall [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 751.153786] env[61868]: DEBUG nova.compute.manager [-] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 751.153883] env[61868]: DEBUG nova.network.neutron [-] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 751.178214] env[61868]: DEBUG nova.network.neutron [-] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 751.522918] env[61868]: DEBUG nova.compute.utils [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 751.524100] env[61868]: DEBUG nova.compute.manager [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 751.524476] env[61868]: DEBUG nova.network.neutron [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 751.608878] env[61868]: DEBUG nova.policy [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cce2af4737904177a83d12587d265dc9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1a5f4d9f9ef3491286397b85ff547f36', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 751.679549] env[61868]: DEBUG nova.network.neutron [-] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.869019] env[61868]: INFO nova.scheduler.client.report [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] Deleted allocations for instance 66c5f446-91e5-4de1-abb6-f06a0a36020d [ 751.988636] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9358a100-c62b-40a9-98ba-889066340bea {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.996540] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62743826-2c29-44d8-b6a5-4a1439472eed {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.032804] env[61868]: DEBUG nova.network.neutron [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Successfully created port: 7192ca4b-ab2e-4d7c-8dec-36372e362ecc {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 752.035340] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f31770e-0d0c-47c9-a4a3-c98752c7d21d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.038208] env[61868]: DEBUG nova.compute.manager [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 752.048424] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a95fdfda-5dcd-4ce1-87f3-b5a13cf18fea {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.064021] env[61868]: DEBUG nova.compute.provider_tree [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 752.186904] env[61868]: INFO nova.compute.manager [-] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Took 1.03 seconds to deallocate network for instance. [ 752.187931] env[61868]: DEBUG nova.compute.claims [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 752.187931] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.380657] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a8dbe924-68e7-4333-9bd5-06e7528f2492 tempest-ServersTestBootFromVolume-2004391495 tempest-ServersTestBootFromVolume-2004391495-project-member] Lock "66c5f446-91e5-4de1-abb6-f06a0a36020d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 140.138s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.569434] env[61868]: DEBUG nova.scheduler.client.report [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 752.883275] env[61868]: DEBUG nova.compute.manager [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 753.047950] env[61868]: DEBUG nova.compute.manager [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 753.074110] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.058s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.074414] env[61868]: ERROR nova.compute.manager [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4d03d50a-044d-4665-9949-cb040fc108ae, please check neutron logs for more information. [ 753.074414] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Traceback (most recent call last): [ 753.074414] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 753.074414] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] self.driver.spawn(context, instance, image_meta, [ 753.074414] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 753.074414] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 753.074414] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 753.074414] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] vm_ref = self.build_virtual_machine(instance, [ 753.074414] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 753.074414] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] vif_infos = vmwarevif.get_vif_info(self._session, [ 753.074414] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 753.075098] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] for vif in network_info: [ 753.075098] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 753.075098] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] return self._sync_wrapper(fn, *args, **kwargs) [ 753.075098] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 753.075098] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] self.wait() [ 753.075098] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 753.075098] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] self[:] = self._gt.wait() [ 753.075098] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 753.075098] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] return self._exit_event.wait() [ 753.075098] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 753.075098] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] current.throw(*self._exc) [ 753.075098] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 753.075098] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] result = function(*args, **kwargs) [ 753.075924] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 753.075924] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] return func(*args, **kwargs) [ 753.075924] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 753.075924] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] raise e [ 753.075924] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 753.075924] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] nwinfo = self.network_api.allocate_for_instance( [ 753.075924] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 753.075924] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] created_port_ids = self._update_ports_for_instance( [ 753.075924] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 753.075924] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] with excutils.save_and_reraise_exception(): [ 753.075924] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.075924] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] self.force_reraise() [ 753.075924] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.077211] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] raise self.value [ 753.077211] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 753.077211] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] updated_port = self._update_port( [ 753.077211] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.077211] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] _ensure_no_port_binding_failure(port) [ 753.077211] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.077211] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] raise exception.PortBindingFailed(port_id=port['id']) [ 753.077211] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] nova.exception.PortBindingFailed: Binding failed for port 4d03d50a-044d-4665-9949-cb040fc108ae, please check neutron logs for more information. [ 753.077211] env[61868]: ERROR nova.compute.manager [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] [ 753.077211] env[61868]: DEBUG nova.compute.utils [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Binding failed for port 4d03d50a-044d-4665-9949-cb040fc108ae, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 753.078305] env[61868]: DEBUG nova.virt.hardware [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 753.078538] env[61868]: DEBUG nova.virt.hardware [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 753.078719] env[61868]: DEBUG nova.virt.hardware [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 753.078887] env[61868]: DEBUG nova.virt.hardware [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 753.079188] env[61868]: DEBUG nova.virt.hardware [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 753.079369] env[61868]: DEBUG nova.virt.hardware [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 753.079586] env[61868]: DEBUG nova.virt.hardware [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 753.080489] env[61868]: DEBUG nova.virt.hardware [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 753.080489] env[61868]: DEBUG nova.virt.hardware [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 753.080489] env[61868]: DEBUG nova.virt.hardware [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 753.080489] env[61868]: DEBUG nova.virt.hardware [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 753.080754] env[61868]: DEBUG oslo_concurrency.lockutils [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.679s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.081996] env[61868]: INFO nova.compute.claims [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] [instance: df927c0c-b742-4451-9439-60ea63287b99] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 753.084867] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-199165f1-9e86-42be-b7c8-9a3a9752fb2d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.088508] env[61868]: DEBUG nova.compute.manager [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Build of instance 97d1d9c9-214f-4573-9efa-3254ea3f2ee0 was re-scheduled: Binding failed for port 4d03d50a-044d-4665-9949-cb040fc108ae, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 753.088508] env[61868]: DEBUG nova.compute.manager [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 753.088667] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] Acquiring lock "refresh_cache-97d1d9c9-214f-4573-9efa-3254ea3f2ee0" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.088707] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] Acquired lock "refresh_cache-97d1d9c9-214f-4573-9efa-3254ea3f2ee0" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.088882] env[61868]: DEBUG nova.network.neutron [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 753.097362] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea73b4b8-cb45-48b0-96d9-d05fc5e44ff1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.332218] env[61868]: ERROR nova.compute.manager [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7192ca4b-ab2e-4d7c-8dec-36372e362ecc, please check neutron logs for more information. [ 753.332218] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 753.332218] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 753.332218] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 753.332218] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 753.332218] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 753.332218] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 753.332218] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 753.332218] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.332218] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 753.332218] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.332218] env[61868]: ERROR nova.compute.manager raise self.value [ 753.332218] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 753.332218] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 753.332218] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.332218] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 753.332773] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.332773] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 753.332773] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7192ca4b-ab2e-4d7c-8dec-36372e362ecc, please check neutron logs for more information. [ 753.332773] env[61868]: ERROR nova.compute.manager [ 753.332773] env[61868]: Traceback (most recent call last): [ 753.332773] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 753.332773] env[61868]: listener.cb(fileno) [ 753.332773] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 753.332773] env[61868]: result = function(*args, **kwargs) [ 753.332773] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 753.332773] env[61868]: return func(*args, **kwargs) [ 753.332773] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 753.332773] env[61868]: raise e [ 753.332773] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 753.332773] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 753.332773] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 753.332773] env[61868]: created_port_ids = self._update_ports_for_instance( [ 753.332773] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 753.332773] env[61868]: with excutils.save_and_reraise_exception(): [ 753.332773] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.332773] env[61868]: self.force_reraise() [ 753.332773] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.332773] env[61868]: raise self.value [ 753.332773] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 753.332773] env[61868]: updated_port = self._update_port( [ 753.332773] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.332773] env[61868]: _ensure_no_port_binding_failure(port) [ 753.332773] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.332773] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 753.333636] env[61868]: nova.exception.PortBindingFailed: Binding failed for port 7192ca4b-ab2e-4d7c-8dec-36372e362ecc, please check neutron logs for more information. [ 753.333636] env[61868]: Removing descriptor: 16 [ 753.333636] env[61868]: ERROR nova.compute.manager [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7192ca4b-ab2e-4d7c-8dec-36372e362ecc, please check neutron logs for more information. [ 753.333636] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Traceback (most recent call last): [ 753.333636] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 753.333636] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] yield resources [ 753.333636] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 753.333636] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] self.driver.spawn(context, instance, image_meta, [ 753.333636] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 753.333636] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 753.333636] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 753.333636] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] vm_ref = self.build_virtual_machine(instance, [ 753.334040] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 753.334040] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] vif_infos = vmwarevif.get_vif_info(self._session, [ 753.334040] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 753.334040] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] for vif in network_info: [ 753.334040] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 753.334040] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] return self._sync_wrapper(fn, *args, **kwargs) [ 753.334040] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 753.334040] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] self.wait() [ 753.334040] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 753.334040] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] self[:] = self._gt.wait() [ 753.334040] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 753.334040] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] return self._exit_event.wait() [ 753.334040] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 753.334501] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] result = hub.switch() [ 753.334501] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 753.334501] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] return self.greenlet.switch() [ 753.334501] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 753.334501] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] result = function(*args, **kwargs) [ 753.334501] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 753.334501] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] return func(*args, **kwargs) [ 753.334501] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 753.334501] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] raise e [ 753.334501] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 753.334501] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] nwinfo = self.network_api.allocate_for_instance( [ 753.334501] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 753.334501] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] created_port_ids = self._update_ports_for_instance( [ 753.334979] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 753.334979] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] with excutils.save_and_reraise_exception(): [ 753.334979] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.334979] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] self.force_reraise() [ 753.334979] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.334979] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] raise self.value [ 753.334979] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 753.334979] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] updated_port = self._update_port( [ 753.334979] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.334979] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] _ensure_no_port_binding_failure(port) [ 753.334979] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.334979] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] raise exception.PortBindingFailed(port_id=port['id']) [ 753.335341] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] nova.exception.PortBindingFailed: Binding failed for port 7192ca4b-ab2e-4d7c-8dec-36372e362ecc, please check neutron logs for more information. [ 753.335341] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] [ 753.335341] env[61868]: INFO nova.compute.manager [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Terminating instance [ 753.339573] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Acquiring lock "refresh_cache-2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.339746] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Acquired lock "refresh_cache-2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.339899] env[61868]: DEBUG nova.network.neutron [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 753.413428] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.615866] env[61868]: DEBUG nova.network.neutron [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 753.708289] env[61868]: DEBUG nova.network.neutron [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.732616] env[61868]: DEBUG nova.compute.manager [req-e7d79144-5555-422d-9963-a7bec526025b req-d8de1ef0-9b41-40ca-9989-eadc5cd71310 service nova] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Received event network-changed-7192ca4b-ab2e-4d7c-8dec-36372e362ecc {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 753.732740] env[61868]: DEBUG nova.compute.manager [req-e7d79144-5555-422d-9963-a7bec526025b req-d8de1ef0-9b41-40ca-9989-eadc5cd71310 service nova] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Refreshing instance network info cache due to event network-changed-7192ca4b-ab2e-4d7c-8dec-36372e362ecc. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 753.732881] env[61868]: DEBUG oslo_concurrency.lockutils [req-e7d79144-5555-422d-9963-a7bec526025b req-d8de1ef0-9b41-40ca-9989-eadc5cd71310 service nova] Acquiring lock "refresh_cache-2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.866299] env[61868]: DEBUG nova.network.neutron [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 753.958539] env[61868]: DEBUG nova.network.neutron [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.215035] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] Releasing lock "refresh_cache-97d1d9c9-214f-4573-9efa-3254ea3f2ee0" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.215988] env[61868]: DEBUG nova.compute.manager [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 754.216354] env[61868]: DEBUG nova.compute.manager [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 754.216625] env[61868]: DEBUG nova.network.neutron [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 754.320940] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Acquiring lock "ca85dfd1-e794-41dc-a337-fbd99998e1fa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.321249] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Lock "ca85dfd1-e794-41dc-a337-fbd99998e1fa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.342127] env[61868]: DEBUG nova.network.neutron [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 754.464947] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Releasing lock "refresh_cache-2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.465453] env[61868]: DEBUG nova.compute.manager [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 754.465715] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 754.466500] env[61868]: DEBUG oslo_concurrency.lockutils [req-e7d79144-5555-422d-9963-a7bec526025b req-d8de1ef0-9b41-40ca-9989-eadc5cd71310 service nova] Acquired lock "refresh_cache-2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.466951] env[61868]: DEBUG nova.network.neutron [req-e7d79144-5555-422d-9963-a7bec526025b req-d8de1ef0-9b41-40ca-9989-eadc5cd71310 service nova] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Refreshing network info cache for port 7192ca4b-ab2e-4d7c-8dec-36372e362ecc {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 754.470830] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6d11cef9-34ff-4fb1-96c7-30df39f87274 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.479070] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e66d4a7d-ea9c-4edb-91b0-ee457b81abfb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.505565] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3 could not be found. [ 754.505950] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 754.506287] env[61868]: INFO nova.compute.manager [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Took 0.04 seconds to destroy the instance on the hypervisor. [ 754.509444] env[61868]: DEBUG oslo.service.loopingcall [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 754.512946] env[61868]: DEBUG nova.compute.manager [-] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 754.512946] env[61868]: DEBUG nova.network.neutron [-] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 754.536719] env[61868]: DEBUG nova.network.neutron [-] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 754.575828] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c46e465-0052-4bbe-b303-5ca5fe2db284 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.584270] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58b8d700-9d34-4a44-b281-67afd735e602 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.617017] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e6077b2-ccb1-44e1-81af-41034e7573d6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.622417] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26e6fb59-e8be-4636-9969-cd6c1bb80f22 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.636477] env[61868]: DEBUG nova.compute.provider_tree [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 754.848489] env[61868]: DEBUG nova.network.neutron [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.993315] env[61868]: DEBUG nova.network.neutron [req-e7d79144-5555-422d-9963-a7bec526025b req-d8de1ef0-9b41-40ca-9989-eadc5cd71310 service nova] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 755.040717] env[61868]: DEBUG nova.network.neutron [-] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.133412] env[61868]: DEBUG nova.network.neutron [req-e7d79144-5555-422d-9963-a7bec526025b req-d8de1ef0-9b41-40ca-9989-eadc5cd71310 service nova] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.138611] env[61868]: DEBUG nova.scheduler.client.report [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 755.351378] env[61868]: INFO nova.compute.manager [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] [instance: 97d1d9c9-214f-4573-9efa-3254ea3f2ee0] Took 1.13 seconds to deallocate network for instance. [ 755.546812] env[61868]: INFO nova.compute.manager [-] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Took 1.04 seconds to deallocate network for instance. [ 755.549647] env[61868]: DEBUG nova.compute.claims [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 755.549830] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.635805] env[61868]: DEBUG oslo_concurrency.lockutils [req-e7d79144-5555-422d-9963-a7bec526025b req-d8de1ef0-9b41-40ca-9989-eadc5cd71310 service nova] Releasing lock "refresh_cache-2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 755.644287] env[61868]: DEBUG oslo_concurrency.lockutils [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.563s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.644287] env[61868]: DEBUG nova.compute.manager [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] [instance: df927c0c-b742-4451-9439-60ea63287b99] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 755.646881] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.599s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.760187] env[61868]: DEBUG nova.compute.manager [req-cd26987e-6cbc-4d86-8964-7f0daf4b53ea req-7f11a167-ddab-435f-845d-1371d6b8cf05 service nova] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Received event network-vif-deleted-7192ca4b-ab2e-4d7c-8dec-36372e362ecc {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 755.972107] env[61868]: DEBUG oslo_concurrency.lockutils [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Acquiring lock "19c2720c-90bc-47f6-999b-6031f893408d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.972107] env[61868]: DEBUG oslo_concurrency.lockutils [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Lock "19c2720c-90bc-47f6-999b-6031f893408d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 756.151775] env[61868]: DEBUG nova.compute.utils [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 756.156345] env[61868]: DEBUG nova.compute.manager [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] [instance: df927c0c-b742-4451-9439-60ea63287b99] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 756.156519] env[61868]: DEBUG nova.network.neutron [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] [instance: df927c0c-b742-4451-9439-60ea63287b99] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 756.211580] env[61868]: DEBUG nova.policy [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '085fb9cd02234e5783f528a4ac292bfe', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4f10b977c1bd4b75a93cfc61751e6ea7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 756.399211] env[61868]: INFO nova.scheduler.client.report [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] Deleted allocations for instance 97d1d9c9-214f-4573-9efa-3254ea3f2ee0 [ 756.562313] env[61868]: DEBUG nova.network.neutron [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] [instance: df927c0c-b742-4451-9439-60ea63287b99] Successfully created port: 12eed284-8b74-4be9-ab30-4d3e30c0d76f {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 756.652539] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-480fe450-e624-4eb8-9b37-8a87af9b7df8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.658955] env[61868]: DEBUG nova.compute.manager [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] [instance: df927c0c-b742-4451-9439-60ea63287b99] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 756.662321] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cebeff5e-23ff-4404-93c0-d341ba0f9314 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.696271] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e566e0d-c330-443b-b639-f9bc447e424a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.705065] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaef67bb-fcc1-4bad-afec-57361c125ab4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.720745] env[61868]: DEBUG nova.compute.provider_tree [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 756.914603] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9562de5a-534b-4f28-93de-8ae4cfc08336 tempest-ServersV294TestFqdnHostnames-862118082 tempest-ServersV294TestFqdnHostnames-862118082-project-member] Lock "97d1d9c9-214f-4573-9efa-3254ea3f2ee0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 134.780s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.225474] env[61868]: DEBUG nova.scheduler.client.report [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 757.418406] env[61868]: DEBUG nova.compute.manager [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 757.598155] env[61868]: ERROR nova.compute.manager [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 12eed284-8b74-4be9-ab30-4d3e30c0d76f, please check neutron logs for more information. [ 757.598155] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 757.598155] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 757.598155] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 757.598155] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 757.598155] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 757.598155] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 757.598155] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 757.598155] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 757.598155] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 757.598155] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 757.598155] env[61868]: ERROR nova.compute.manager raise self.value [ 757.598155] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 757.598155] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 757.598155] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 757.598155] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 757.598642] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 757.598642] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 757.598642] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 12eed284-8b74-4be9-ab30-4d3e30c0d76f, please check neutron logs for more information. [ 757.598642] env[61868]: ERROR nova.compute.manager [ 757.598642] env[61868]: Traceback (most recent call last): [ 757.598642] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 757.598642] env[61868]: listener.cb(fileno) [ 757.598642] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 757.598642] env[61868]: result = function(*args, **kwargs) [ 757.598642] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 757.598642] env[61868]: return func(*args, **kwargs) [ 757.598642] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 757.598642] env[61868]: raise e [ 757.598642] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 757.598642] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 757.598642] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 757.598642] env[61868]: created_port_ids = self._update_ports_for_instance( [ 757.598642] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 757.598642] env[61868]: with excutils.save_and_reraise_exception(): [ 757.598642] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 757.598642] env[61868]: self.force_reraise() [ 757.598642] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 757.598642] env[61868]: raise self.value [ 757.598642] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 757.598642] env[61868]: updated_port = self._update_port( [ 757.598642] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 757.598642] env[61868]: _ensure_no_port_binding_failure(port) [ 757.598642] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 757.598642] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 757.599639] env[61868]: nova.exception.PortBindingFailed: Binding failed for port 12eed284-8b74-4be9-ab30-4d3e30c0d76f, please check neutron logs for more information. [ 757.599639] env[61868]: Removing descriptor: 16 [ 757.673120] env[61868]: DEBUG nova.compute.manager [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] [instance: df927c0c-b742-4451-9439-60ea63287b99] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 757.699462] env[61868]: DEBUG nova.virt.hardware [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 757.699714] env[61868]: DEBUG nova.virt.hardware [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 757.699870] env[61868]: DEBUG nova.virt.hardware [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 757.700057] env[61868]: DEBUG nova.virt.hardware [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 757.700271] env[61868]: DEBUG nova.virt.hardware [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 757.700358] env[61868]: DEBUG nova.virt.hardware [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 757.700560] env[61868]: DEBUG nova.virt.hardware [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 757.700715] env[61868]: DEBUG nova.virt.hardware [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 757.700874] env[61868]: DEBUG nova.virt.hardware [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 757.701821] env[61868]: DEBUG nova.virt.hardware [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 757.702209] env[61868]: DEBUG nova.virt.hardware [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 757.703246] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91bde97f-61d3-449c-a641-ab83d4a88d72 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.713099] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d45be5a9-9dbd-438b-baa3-7b0a3b5fe80b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.728776] env[61868]: ERROR nova.compute.manager [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] [instance: df927c0c-b742-4451-9439-60ea63287b99] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 12eed284-8b74-4be9-ab30-4d3e30c0d76f, please check neutron logs for more information. [ 757.728776] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] Traceback (most recent call last): [ 757.728776] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 757.728776] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] yield resources [ 757.728776] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 757.728776] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] self.driver.spawn(context, instance, image_meta, [ 757.728776] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 757.728776] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] self._vmops.spawn(context, instance, image_meta, injected_files, [ 757.728776] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 757.728776] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] vm_ref = self.build_virtual_machine(instance, [ 757.728776] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 757.729160] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] vif_infos = vmwarevif.get_vif_info(self._session, [ 757.729160] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 757.729160] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] for vif in network_info: [ 757.729160] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 757.729160] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] return self._sync_wrapper(fn, *args, **kwargs) [ 757.729160] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 757.729160] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] self.wait() [ 757.729160] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 757.729160] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] self[:] = self._gt.wait() [ 757.729160] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 757.729160] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] return self._exit_event.wait() [ 757.729160] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 757.729160] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] current.throw(*self._exc) [ 757.729603] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 757.729603] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] result = function(*args, **kwargs) [ 757.729603] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 757.729603] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] return func(*args, **kwargs) [ 757.729603] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 757.729603] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] raise e [ 757.729603] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 757.729603] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] nwinfo = self.network_api.allocate_for_instance( [ 757.729603] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 757.729603] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] created_port_ids = self._update_ports_for_instance( [ 757.729603] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 757.729603] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] with excutils.save_and_reraise_exception(): [ 757.729603] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 757.730056] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] self.force_reraise() [ 757.730056] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 757.730056] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] raise self.value [ 757.730056] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 757.730056] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] updated_port = self._update_port( [ 757.730056] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 757.730056] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] _ensure_no_port_binding_failure(port) [ 757.730056] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 757.730056] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] raise exception.PortBindingFailed(port_id=port['id']) [ 757.730056] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] nova.exception.PortBindingFailed: Binding failed for port 12eed284-8b74-4be9-ab30-4d3e30c0d76f, please check neutron logs for more information. [ 757.730056] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] [ 757.735196] env[61868]: INFO nova.compute.manager [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] [instance: df927c0c-b742-4451-9439-60ea63287b99] Terminating instance [ 757.735465] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.089s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.736041] env[61868]: ERROR nova.compute.manager [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port df04fac4-ceb3-44ad-94b4-c043fbd3868c, please check neutron logs for more information. [ 757.736041] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Traceback (most recent call last): [ 757.736041] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 757.736041] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] self.driver.spawn(context, instance, image_meta, [ 757.736041] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 757.736041] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] self._vmops.spawn(context, instance, image_meta, injected_files, [ 757.736041] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 757.736041] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] vm_ref = self.build_virtual_machine(instance, [ 757.736041] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 757.736041] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] vif_infos = vmwarevif.get_vif_info(self._session, [ 757.736041] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 757.736393] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] for vif in network_info: [ 757.736393] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 757.736393] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] return self._sync_wrapper(fn, *args, **kwargs) [ 757.736393] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 757.736393] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] self.wait() [ 757.736393] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 757.736393] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] self[:] = self._gt.wait() [ 757.736393] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 757.736393] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] return self._exit_event.wait() [ 757.736393] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 757.736393] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] current.throw(*self._exc) [ 757.736393] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 757.736393] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] result = function(*args, **kwargs) [ 757.736765] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 757.736765] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] return func(*args, **kwargs) [ 757.736765] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 757.736765] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] raise e [ 757.736765] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 757.736765] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] nwinfo = self.network_api.allocate_for_instance( [ 757.736765] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 757.736765] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] created_port_ids = self._update_ports_for_instance( [ 757.736765] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 757.736765] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] with excutils.save_and_reraise_exception(): [ 757.736765] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 757.736765] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] self.force_reraise() [ 757.736765] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 757.737174] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] raise self.value [ 757.737174] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 757.737174] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] updated_port = self._update_port( [ 757.737174] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 757.737174] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] _ensure_no_port_binding_failure(port) [ 757.737174] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 757.737174] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] raise exception.PortBindingFailed(port_id=port['id']) [ 757.737174] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] nova.exception.PortBindingFailed: Binding failed for port df04fac4-ceb3-44ad-94b4-c043fbd3868c, please check neutron logs for more information. [ 757.737174] env[61868]: ERROR nova.compute.manager [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] [ 757.737174] env[61868]: DEBUG nova.compute.utils [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Binding failed for port df04fac4-ceb3-44ad-94b4-c043fbd3868c, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 757.737905] env[61868]: DEBUG oslo_concurrency.lockutils [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] Acquiring lock "refresh_cache-df927c0c-b742-4451-9439-60ea63287b99" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.738484] env[61868]: DEBUG oslo_concurrency.lockutils [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] Acquired lock "refresh_cache-df927c0c-b742-4451-9439-60ea63287b99" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.738724] env[61868]: DEBUG nova.network.neutron [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] [instance: df927c0c-b742-4451-9439-60ea63287b99] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 757.739726] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.750s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.741192] env[61868]: INFO nova.compute.claims [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 757.745602] env[61868]: DEBUG nova.compute.manager [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Build of instance 84dc8092-c24a-4587-bfcb-865b6eb129aa was re-scheduled: Binding failed for port df04fac4-ceb3-44ad-94b4-c043fbd3868c, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 757.745602] env[61868]: DEBUG nova.compute.manager [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 757.745602] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Acquiring lock "refresh_cache-84dc8092-c24a-4587-bfcb-865b6eb129aa" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.745602] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Acquired lock "refresh_cache-84dc8092-c24a-4587-bfcb-865b6eb129aa" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.745802] env[61868]: DEBUG nova.network.neutron [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 757.865948] env[61868]: DEBUG nova.compute.manager [req-a6abec1c-88b4-4467-8d51-ef56d482ac88 req-62e5f4a8-9598-44c2-9737-c911e456dd73 service nova] [instance: df927c0c-b742-4451-9439-60ea63287b99] Received event network-changed-12eed284-8b74-4be9-ab30-4d3e30c0d76f {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 757.865948] env[61868]: DEBUG nova.compute.manager [req-a6abec1c-88b4-4467-8d51-ef56d482ac88 req-62e5f4a8-9598-44c2-9737-c911e456dd73 service nova] [instance: df927c0c-b742-4451-9439-60ea63287b99] Refreshing instance network info cache due to event network-changed-12eed284-8b74-4be9-ab30-4d3e30c0d76f. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 757.865948] env[61868]: DEBUG oslo_concurrency.lockutils [req-a6abec1c-88b4-4467-8d51-ef56d482ac88 req-62e5f4a8-9598-44c2-9737-c911e456dd73 service nova] Acquiring lock "refresh_cache-df927c0c-b742-4451-9439-60ea63287b99" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.944257] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.279017] env[61868]: DEBUG nova.network.neutron [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] [instance: df927c0c-b742-4451-9439-60ea63287b99] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 758.281551] env[61868]: DEBUG nova.network.neutron [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 758.472902] env[61868]: DEBUG nova.network.neutron [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] [instance: df927c0c-b742-4451-9439-60ea63287b99] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.478762] env[61868]: DEBUG nova.network.neutron [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.976459] env[61868]: DEBUG oslo_concurrency.lockutils [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] Releasing lock "refresh_cache-df927c0c-b742-4451-9439-60ea63287b99" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.977955] env[61868]: DEBUG nova.compute.manager [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] [instance: df927c0c-b742-4451-9439-60ea63287b99] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 758.977955] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] [instance: df927c0c-b742-4451-9439-60ea63287b99] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 758.977955] env[61868]: DEBUG oslo_concurrency.lockutils [req-a6abec1c-88b4-4467-8d51-ef56d482ac88 req-62e5f4a8-9598-44c2-9737-c911e456dd73 service nova] Acquired lock "refresh_cache-df927c0c-b742-4451-9439-60ea63287b99" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.978654] env[61868]: DEBUG nova.network.neutron [req-a6abec1c-88b4-4467-8d51-ef56d482ac88 req-62e5f4a8-9598-44c2-9737-c911e456dd73 service nova] [instance: df927c0c-b742-4451-9439-60ea63287b99] Refreshing network info cache for port 12eed284-8b74-4be9-ab30-4d3e30c0d76f {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 758.983806] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7217c66c-9a38-4882-bfd4-1e9b4f1548e4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.990821] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Releasing lock "refresh_cache-84dc8092-c24a-4587-bfcb-865b6eb129aa" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.991298] env[61868]: DEBUG nova.compute.manager [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 758.991298] env[61868]: DEBUG nova.compute.manager [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 758.991519] env[61868]: DEBUG nova.network.neutron [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 759.003814] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38aafdac-1be7-40d8-b3cb-97ced1c79f9f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.018855] env[61868]: DEBUG nova.network.neutron [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 759.030800] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] [instance: df927c0c-b742-4451-9439-60ea63287b99] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance df927c0c-b742-4451-9439-60ea63287b99 could not be found. [ 759.031123] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] [instance: df927c0c-b742-4451-9439-60ea63287b99] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 759.031402] env[61868]: INFO nova.compute.manager [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] [instance: df927c0c-b742-4451-9439-60ea63287b99] Took 0.05 seconds to destroy the instance on the hypervisor. [ 759.031703] env[61868]: DEBUG oslo.service.loopingcall [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 759.035040] env[61868]: DEBUG nova.compute.manager [-] [instance: df927c0c-b742-4451-9439-60ea63287b99] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 759.035040] env[61868]: DEBUG nova.network.neutron [-] [instance: df927c0c-b742-4451-9439-60ea63287b99] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 759.059364] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 759.059364] env[61868]: DEBUG nova.network.neutron [-] [instance: df927c0c-b742-4451-9439-60ea63287b99] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 759.063553] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 759.199770] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9eb8c2a-cada-4559-ae72-259a0bef84b8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.208787] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35870726-fc5b-4334-8314-555099c009b6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.244630] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-897dd39f-f927-4520-a0b0-7cbad6537881 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.252178] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93a36f6d-903d-47c6-9486-10fa2e48260c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.267206] env[61868]: DEBUG nova.compute.provider_tree [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 759.513546] env[61868]: DEBUG nova.network.neutron [req-a6abec1c-88b4-4467-8d51-ef56d482ac88 req-62e5f4a8-9598-44c2-9737-c911e456dd73 service nova] [instance: df927c0c-b742-4451-9439-60ea63287b99] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 759.522739] env[61868]: DEBUG nova.network.neutron [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.563785] env[61868]: DEBUG nova.network.neutron [-] [instance: df927c0c-b742-4451-9439-60ea63287b99] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.574021] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 759.574021] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Starting heal instance info cache {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 759.574021] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Rebuilding the list of instances to heal {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 759.639110] env[61868]: DEBUG nova.network.neutron [req-a6abec1c-88b4-4467-8d51-ef56d482ac88 req-62e5f4a8-9598-44c2-9737-c911e456dd73 service nova] [instance: df927c0c-b742-4451-9439-60ea63287b99] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.770711] env[61868]: DEBUG nova.scheduler.client.report [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 760.030369] env[61868]: INFO nova.compute.manager [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 84dc8092-c24a-4587-bfcb-865b6eb129aa] Took 1.04 seconds to deallocate network for instance. [ 760.070995] env[61868]: INFO nova.compute.manager [-] [instance: df927c0c-b742-4451-9439-60ea63287b99] Took 1.04 seconds to deallocate network for instance. [ 760.077019] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Skipping network cache update for instance because it is Building. {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 760.077019] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Skipping network cache update for instance because it is Building. {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 760.077019] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: df927c0c-b742-4451-9439-60ea63287b99] Skipping network cache update for instance because it is Building. {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 760.077019] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Skipping network cache update for instance because it is Building. {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 760.077019] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Didn't find any instances for network info cache update. {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 760.077019] env[61868]: DEBUG nova.compute.claims [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] [instance: df927c0c-b742-4451-9439-60ea63287b99] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 760.077396] env[61868]: DEBUG oslo_concurrency.lockutils [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.077396] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 760.077530] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 760.077798] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 760.078446] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 760.078882] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 760.079267] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 760.080133] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61868) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 760.082414] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager.update_available_resource {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 760.145698] env[61868]: DEBUG oslo_concurrency.lockutils [req-a6abec1c-88b4-4467-8d51-ef56d482ac88 req-62e5f4a8-9598-44c2-9737-c911e456dd73 service nova] Releasing lock "refresh_cache-df927c0c-b742-4451-9439-60ea63287b99" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.145994] env[61868]: DEBUG nova.compute.manager [req-a6abec1c-88b4-4467-8d51-ef56d482ac88 req-62e5f4a8-9598-44c2-9737-c911e456dd73 service nova] [instance: df927c0c-b742-4451-9439-60ea63287b99] Received event network-vif-deleted-12eed284-8b74-4be9-ab30-4d3e30c0d76f {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 760.276458] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.537s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.277240] env[61868]: DEBUG nova.compute.manager [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 760.280679] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.348s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.283561] env[61868]: INFO nova.compute.claims [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 760.585607] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.606154] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquiring lock "75fdbc11-1ef3-442d-bcf4-55069ba9ab62" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.606154] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lock "75fdbc11-1ef3-442d-bcf4-55069ba9ab62" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.788361] env[61868]: DEBUG nova.compute.utils [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 760.789793] env[61868]: DEBUG nova.compute.manager [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Not allocating networking since 'none' was specified. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 761.069210] env[61868]: INFO nova.scheduler.client.report [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Deleted allocations for instance 84dc8092-c24a-4587-bfcb-865b6eb129aa [ 761.293303] env[61868]: DEBUG nova.compute.manager [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 761.578916] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0ff58eb3-714e-4b0f-b4b5-4dc2c38cda56 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Lock "84dc8092-c24a-4587-bfcb-865b6eb129aa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 138.391s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.650954] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eae709c-ad57-42a5-9147-3b59928c63c2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.660273] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-798235c8-9224-4a94-8f98-a488211405ca {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.692035] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-334e315e-cbaa-4988-a600-531d8977c04e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.699806] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33d00ec5-6d30-4109-8203-58454dadc5d9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.712914] env[61868]: DEBUG nova.compute.provider_tree [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 762.085205] env[61868]: DEBUG nova.compute.manager [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 762.216852] env[61868]: DEBUG nova.scheduler.client.report [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 762.306512] env[61868]: DEBUG nova.compute.manager [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 762.337585] env[61868]: DEBUG nova.virt.hardware [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 762.337852] env[61868]: DEBUG nova.virt.hardware [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 762.338012] env[61868]: DEBUG nova.virt.hardware [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 762.338205] env[61868]: DEBUG nova.virt.hardware [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 762.338351] env[61868]: DEBUG nova.virt.hardware [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 762.338497] env[61868]: DEBUG nova.virt.hardware [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 762.338703] env[61868]: DEBUG nova.virt.hardware [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 762.338859] env[61868]: DEBUG nova.virt.hardware [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 762.340093] env[61868]: DEBUG nova.virt.hardware [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 762.342067] env[61868]: DEBUG nova.virt.hardware [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 762.342310] env[61868]: DEBUG nova.virt.hardware [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 762.343908] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cc64d84-f458-4d5c-9ae2-c7e122d15a73 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.352011] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45f88dbc-4792-49f0-a93e-96981879502a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.367431] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Instance VIF info [] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 762.373609] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Creating folder: Project (628fdd10207f4e94b1c69de3f726f3e2). Parent ref: group-v281478. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 762.373900] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-43009131-1ed4-49c3-9b19-12df337ee4d7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.382973] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Created folder: Project (628fdd10207f4e94b1c69de3f726f3e2) in parent group-v281478. [ 762.383168] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Creating folder: Instances. Parent ref: group-v281495. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 762.383442] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fb94587a-a43c-4fab-91ae-198691f90930 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.392431] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Created folder: Instances in parent group-v281495. [ 762.392861] env[61868]: DEBUG oslo.service.loopingcall [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 762.393069] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 762.393266] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6030b90e-05e4-45f5-8e89-53ab5fffc1af {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.410032] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 762.410032] env[61868]: value = "task-1315321" [ 762.410032] env[61868]: _type = "Task" [ 762.410032] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.419837] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315321, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.606594] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.721401] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.441s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.721944] env[61868]: DEBUG nova.compute.manager [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 762.725032] env[61868]: DEBUG oslo_concurrency.lockutils [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.848s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.726498] env[61868]: INFO nova.compute.claims [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 762.922233] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315321, 'name': CreateVM_Task, 'duration_secs': 0.23218} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.922233] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 762.922233] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.922233] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.922233] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 762.922233] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-406fa562-0fcc-4104-9ec6-362a0293d01c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.926248] env[61868]: DEBUG oslo_vmware.api [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Waiting for the task: (returnval){ [ 762.926248] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52b029e0-6eb2-a79f-f4fa-c8270c5e57fc" [ 762.926248] env[61868]: _type = "Task" [ 762.926248] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.934152] env[61868]: DEBUG oslo_vmware.api [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52b029e0-6eb2-a79f-f4fa-c8270c5e57fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.234692] env[61868]: DEBUG nova.compute.utils [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 763.234692] env[61868]: DEBUG nova.compute.manager [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 763.234692] env[61868]: DEBUG nova.network.neutron [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 763.285751] env[61868]: DEBUG nova.policy [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '50e162441b7c442095a1b720951f5702', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '75088c6912e1464ca00dcd4daaef7b3e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 763.436537] env[61868]: DEBUG oslo_vmware.api [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52b029e0-6eb2-a79f-f4fa-c8270c5e57fc, 'name': SearchDatastore_Task, 'duration_secs': 0.009476} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.436844] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.437089] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 763.437322] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.437463] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.437637] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 763.437882] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cbfc8730-8ac6-40be-82fe-0b5bc9587494 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.445594] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 763.445764] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 763.446466] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe6a3c17-4458-4221-9cb8-dbad9a57eec5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.451443] env[61868]: DEBUG oslo_vmware.api [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Waiting for the task: (returnval){ [ 763.451443] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]525b735a-2e39-b225-1035-97e3b859ac1c" [ 763.451443] env[61868]: _type = "Task" [ 763.451443] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.459307] env[61868]: DEBUG oslo_vmware.api [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]525b735a-2e39-b225-1035-97e3b859ac1c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.571749] env[61868]: DEBUG nova.network.neutron [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Successfully created port: 1ef56e50-9ea3-48fb-b301-847f27e579b0 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 763.659484] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Acquiring lock "c6f99cbc-3e81-44b4-b3d1-d8cd821829f2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.659717] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Lock "c6f99cbc-3e81-44b4-b3d1-d8cd821829f2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 763.737844] env[61868]: DEBUG nova.compute.manager [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 763.968490] env[61868]: DEBUG oslo_vmware.api [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]525b735a-2e39-b225-1035-97e3b859ac1c, 'name': SearchDatastore_Task, 'duration_secs': 0.008394} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.972443] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a6aa0f7-fb37-4b8b-bfe0-f7f94450200a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.978187] env[61868]: DEBUG oslo_vmware.api [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Waiting for the task: (returnval){ [ 763.978187] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52add077-a67e-2c72-f5ce-374c8ead2b8f" [ 763.978187] env[61868]: _type = "Task" [ 763.978187] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.987590] env[61868]: DEBUG oslo_vmware.api [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52add077-a67e-2c72-f5ce-374c8ead2b8f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.153700] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f634a1c9-7001-4bf5-b5bc-4f94677d926d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.163834] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3783ae87-1c10-4e64-9915-50bb587aa563 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.201277] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15a01de6-c1d0-4fee-b733-d12562da56ae {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.208682] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bda33a56-2bab-4c71-8200-d051f775e71a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.225438] env[61868]: DEBUG nova.compute.provider_tree [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 764.303923] env[61868]: DEBUG nova.compute.manager [req-1b4a9d65-4450-456b-9b8e-58f5c8871409 req-c6579a96-40b7-46f7-8f58-9cee45ca8556 service nova] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Received event network-changed-1ef56e50-9ea3-48fb-b301-847f27e579b0 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 764.304312] env[61868]: DEBUG nova.compute.manager [req-1b4a9d65-4450-456b-9b8e-58f5c8871409 req-c6579a96-40b7-46f7-8f58-9cee45ca8556 service nova] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Refreshing instance network info cache due to event network-changed-1ef56e50-9ea3-48fb-b301-847f27e579b0. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 764.304414] env[61868]: DEBUG oslo_concurrency.lockutils [req-1b4a9d65-4450-456b-9b8e-58f5c8871409 req-c6579a96-40b7-46f7-8f58-9cee45ca8556 service nova] Acquiring lock "refresh_cache-a0fda7cf-21b6-4a22-bf56-09e848b4dc9e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.304714] env[61868]: DEBUG oslo_concurrency.lockutils [req-1b4a9d65-4450-456b-9b8e-58f5c8871409 req-c6579a96-40b7-46f7-8f58-9cee45ca8556 service nova] Acquired lock "refresh_cache-a0fda7cf-21b6-4a22-bf56-09e848b4dc9e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.304714] env[61868]: DEBUG nova.network.neutron [req-1b4a9d65-4450-456b-9b8e-58f5c8871409 req-c6579a96-40b7-46f7-8f58-9cee45ca8556 service nova] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Refreshing network info cache for port 1ef56e50-9ea3-48fb-b301-847f27e579b0 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 764.477270] env[61868]: ERROR nova.compute.manager [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1ef56e50-9ea3-48fb-b301-847f27e579b0, please check neutron logs for more information. [ 764.477270] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 764.477270] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 764.477270] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 764.477270] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 764.477270] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 764.477270] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 764.477270] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 764.477270] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 764.477270] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 764.477270] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 764.477270] env[61868]: ERROR nova.compute.manager raise self.value [ 764.477270] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 764.477270] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 764.477270] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 764.477270] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 764.477883] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 764.477883] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 764.477883] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1ef56e50-9ea3-48fb-b301-847f27e579b0, please check neutron logs for more information. [ 764.477883] env[61868]: ERROR nova.compute.manager [ 764.477883] env[61868]: Traceback (most recent call last): [ 764.477883] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 764.477883] env[61868]: listener.cb(fileno) [ 764.477883] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 764.477883] env[61868]: result = function(*args, **kwargs) [ 764.477883] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 764.477883] env[61868]: return func(*args, **kwargs) [ 764.477883] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 764.477883] env[61868]: raise e [ 764.477883] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 764.477883] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 764.477883] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 764.477883] env[61868]: created_port_ids = self._update_ports_for_instance( [ 764.477883] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 764.477883] env[61868]: with excutils.save_and_reraise_exception(): [ 764.477883] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 764.477883] env[61868]: self.force_reraise() [ 764.477883] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 764.477883] env[61868]: raise self.value [ 764.477883] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 764.477883] env[61868]: updated_port = self._update_port( [ 764.477883] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 764.477883] env[61868]: _ensure_no_port_binding_failure(port) [ 764.477883] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 764.477883] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 764.478821] env[61868]: nova.exception.PortBindingFailed: Binding failed for port 1ef56e50-9ea3-48fb-b301-847f27e579b0, please check neutron logs for more information. [ 764.478821] env[61868]: Removing descriptor: 15 [ 764.488615] env[61868]: DEBUG oslo_vmware.api [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52add077-a67e-2c72-f5ce-374c8ead2b8f, 'name': SearchDatastore_Task, 'duration_secs': 0.008801} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.488877] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.489162] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 4de32ee6-e205-435b-a75e-94c3121048e3/4de32ee6-e205-435b-a75e-94c3121048e3.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 764.489551] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6fbedfd3-357d-495f-84c9-d23acb8f5905 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.496170] env[61868]: DEBUG oslo_vmware.api [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Waiting for the task: (returnval){ [ 764.496170] env[61868]: value = "task-1315322" [ 764.496170] env[61868]: _type = "Task" [ 764.496170] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.505470] env[61868]: DEBUG oslo_vmware.api [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': task-1315322, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.730475] env[61868]: DEBUG nova.scheduler.client.report [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 764.746824] env[61868]: DEBUG nova.compute.manager [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 764.774809] env[61868]: DEBUG nova.virt.hardware [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 764.775125] env[61868]: DEBUG nova.virt.hardware [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 764.775343] env[61868]: DEBUG nova.virt.hardware [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 764.775587] env[61868]: DEBUG nova.virt.hardware [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 764.775764] env[61868]: DEBUG nova.virt.hardware [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 764.775928] env[61868]: DEBUG nova.virt.hardware [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 764.776137] env[61868]: DEBUG nova.virt.hardware [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 764.776345] env[61868]: DEBUG nova.virt.hardware [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 764.776533] env[61868]: DEBUG nova.virt.hardware [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 764.776696] env[61868]: DEBUG nova.virt.hardware [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 764.776866] env[61868]: DEBUG nova.virt.hardware [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 764.777763] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-160ee576-bb38-4888-85e4-a14a99bfc774 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.787962] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08a9c920-36ba-4820-94ce-c9314f83bece {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.802582] env[61868]: ERROR nova.compute.manager [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1ef56e50-9ea3-48fb-b301-847f27e579b0, please check neutron logs for more information. [ 764.802582] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Traceback (most recent call last): [ 764.802582] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 764.802582] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] yield resources [ 764.802582] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 764.802582] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] self.driver.spawn(context, instance, image_meta, [ 764.802582] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 764.802582] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 764.802582] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 764.802582] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] vm_ref = self.build_virtual_machine(instance, [ 764.802582] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 764.802956] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] vif_infos = vmwarevif.get_vif_info(self._session, [ 764.802956] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 764.802956] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] for vif in network_info: [ 764.802956] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 764.802956] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] return self._sync_wrapper(fn, *args, **kwargs) [ 764.802956] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 764.802956] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] self.wait() [ 764.802956] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 764.802956] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] self[:] = self._gt.wait() [ 764.802956] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 764.802956] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] return self._exit_event.wait() [ 764.802956] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 764.802956] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] current.throw(*self._exc) [ 764.803356] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 764.803356] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] result = function(*args, **kwargs) [ 764.803356] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 764.803356] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] return func(*args, **kwargs) [ 764.803356] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 764.803356] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] raise e [ 764.803356] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 764.803356] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] nwinfo = self.network_api.allocate_for_instance( [ 764.803356] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 764.803356] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] created_port_ids = self._update_ports_for_instance( [ 764.803356] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 764.803356] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] with excutils.save_and_reraise_exception(): [ 764.803356] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 764.803717] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] self.force_reraise() [ 764.803717] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 764.803717] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] raise self.value [ 764.803717] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 764.803717] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] updated_port = self._update_port( [ 764.803717] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 764.803717] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] _ensure_no_port_binding_failure(port) [ 764.803717] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 764.803717] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] raise exception.PortBindingFailed(port_id=port['id']) [ 764.803717] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] nova.exception.PortBindingFailed: Binding failed for port 1ef56e50-9ea3-48fb-b301-847f27e579b0, please check neutron logs for more information. [ 764.803717] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] [ 764.803717] env[61868]: INFO nova.compute.manager [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Terminating instance [ 764.804947] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] Acquiring lock "refresh_cache-a0fda7cf-21b6-4a22-bf56-09e848b4dc9e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.825841] env[61868]: DEBUG nova.network.neutron [req-1b4a9d65-4450-456b-9b8e-58f5c8871409 req-c6579a96-40b7-46f7-8f58-9cee45ca8556 service nova] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 764.908191] env[61868]: DEBUG nova.network.neutron [req-1b4a9d65-4450-456b-9b8e-58f5c8871409 req-c6579a96-40b7-46f7-8f58-9cee45ca8556 service nova] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.006302] env[61868]: DEBUG oslo_vmware.api [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': task-1315322, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.490337} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.006604] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 4de32ee6-e205-435b-a75e-94c3121048e3/4de32ee6-e205-435b-a75e-94c3121048e3.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 765.006846] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 765.007010] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2d0b4e86-b0ea-4481-91c7-b39d84f294b5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.014061] env[61868]: DEBUG oslo_vmware.api [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Waiting for the task: (returnval){ [ 765.014061] env[61868]: value = "task-1315323" [ 765.014061] env[61868]: _type = "Task" [ 765.014061] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.020654] env[61868]: DEBUG oslo_vmware.api [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': task-1315323, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.236944] env[61868]: DEBUG oslo_concurrency.lockutils [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.512s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.237634] env[61868]: DEBUG nova.compute.manager [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 765.240233] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.426s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.241640] env[61868]: INFO nova.compute.claims [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 765.410879] env[61868]: DEBUG oslo_concurrency.lockutils [req-1b4a9d65-4450-456b-9b8e-58f5c8871409 req-c6579a96-40b7-46f7-8f58-9cee45ca8556 service nova] Releasing lock "refresh_cache-a0fda7cf-21b6-4a22-bf56-09e848b4dc9e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.411376] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] Acquired lock "refresh_cache-a0fda7cf-21b6-4a22-bf56-09e848b4dc9e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.411563] env[61868]: DEBUG nova.network.neutron [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 765.522831] env[61868]: DEBUG oslo_vmware.api [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': task-1315323, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060606} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.523129] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 765.523867] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c77378bd-cce0-4cba-aa71-abd5bd22c733 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.542637] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Reconfiguring VM instance instance-0000002c to attach disk [datastore1] 4de32ee6-e205-435b-a75e-94c3121048e3/4de32ee6-e205-435b-a75e-94c3121048e3.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 765.542891] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-39da4838-9a6b-4af0-9f8c-81ae8626a880 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.563028] env[61868]: DEBUG oslo_vmware.api [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Waiting for the task: (returnval){ [ 765.563028] env[61868]: value = "task-1315324" [ 765.563028] env[61868]: _type = "Task" [ 765.563028] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.570133] env[61868]: DEBUG oslo_vmware.api [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': task-1315324, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.746740] env[61868]: DEBUG nova.compute.utils [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 765.752365] env[61868]: DEBUG nova.compute.manager [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 765.752365] env[61868]: DEBUG nova.network.neutron [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 765.799140] env[61868]: DEBUG nova.policy [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c14b4ff01717495ca97c7f35f91c2995', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd7fc76299baf4a97b57139e5f1caa16e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 765.931256] env[61868]: DEBUG nova.network.neutron [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 766.055445] env[61868]: DEBUG nova.network.neutron [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.074552] env[61868]: DEBUG oslo_vmware.api [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': task-1315324, 'name': ReconfigVM_Task, 'duration_secs': 0.259359} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.074828] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Reconfigured VM instance instance-0000002c to attach disk [datastore1] 4de32ee6-e205-435b-a75e-94c3121048e3/4de32ee6-e205-435b-a75e-94c3121048e3.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 766.075448] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-09f7ac44-708d-4840-88b6-dd05ab1627fb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.081281] env[61868]: DEBUG oslo_vmware.api [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Waiting for the task: (returnval){ [ 766.081281] env[61868]: value = "task-1315325" [ 766.081281] env[61868]: _type = "Task" [ 766.081281] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.090340] env[61868]: DEBUG oslo_vmware.api [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': task-1315325, 'name': Rename_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.139627] env[61868]: DEBUG nova.network.neutron [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Successfully created port: a9b0a8d5-d2f4-4b81-952f-f194085c3b38 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 766.251054] env[61868]: DEBUG nova.compute.manager [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 766.331913] env[61868]: DEBUG nova.compute.manager [req-7a8e0652-d195-4908-ba63-07b28b4b444d req-5c144db1-dd44-4941-9287-66303ea5171f service nova] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Received event network-vif-deleted-1ef56e50-9ea3-48fb-b301-847f27e579b0 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 766.562965] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] Releasing lock "refresh_cache-a0fda7cf-21b6-4a22-bf56-09e848b4dc9e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.563404] env[61868]: DEBUG nova.compute.manager [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 766.563597] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 766.564644] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c2dc8f76-3c1d-409e-8c08-5b212f2053d0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.574486] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a41fa12d-6ff7-442f-a388-ca945b47729e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.597887] env[61868]: DEBUG oslo_vmware.api [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': task-1315325, 'name': Rename_Task, 'duration_secs': 0.202857} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.600605] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 766.604858] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ef653878-d1db-4334-89f5-eb474b5076fd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.607622] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a0fda7cf-21b6-4a22-bf56-09e848b4dc9e could not be found. [ 766.607622] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 766.607622] env[61868]: INFO nova.compute.manager [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 766.607622] env[61868]: DEBUG oslo.service.loopingcall [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 766.607622] env[61868]: DEBUG nova.compute.manager [-] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 766.607622] env[61868]: DEBUG nova.network.neutron [-] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 766.613852] env[61868]: DEBUG oslo_vmware.api [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Waiting for the task: (returnval){ [ 766.613852] env[61868]: value = "task-1315326" [ 766.613852] env[61868]: _type = "Task" [ 766.613852] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.621201] env[61868]: DEBUG oslo_vmware.api [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': task-1315326, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.627463] env[61868]: DEBUG nova.network.neutron [-] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 766.653644] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4746b0f9-a1f1-442e-adf3-552ce81d6f36 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.662658] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb21f7d4-a75e-4365-a42a-840b1d9cf77a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.694143] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58420312-1679-4a4b-8907-6c8dd6a6f4d0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.702049] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-055ef463-2be2-46e0-ad9a-2ebe94c0ef69 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.715384] env[61868]: DEBUG nova.compute.provider_tree [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Updating inventory in ProviderTree for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 767.041135] env[61868]: ERROR nova.compute.manager [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a9b0a8d5-d2f4-4b81-952f-f194085c3b38, please check neutron logs for more information. [ 767.041135] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 767.041135] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 767.041135] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 767.041135] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 767.041135] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 767.041135] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 767.041135] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 767.041135] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 767.041135] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 767.041135] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 767.041135] env[61868]: ERROR nova.compute.manager raise self.value [ 767.041135] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 767.041135] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 767.041135] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 767.041135] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 767.042147] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 767.042147] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 767.042147] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a9b0a8d5-d2f4-4b81-952f-f194085c3b38, please check neutron logs for more information. [ 767.042147] env[61868]: ERROR nova.compute.manager [ 767.042147] env[61868]: Traceback (most recent call last): [ 767.042147] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 767.042147] env[61868]: listener.cb(fileno) [ 767.042147] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 767.042147] env[61868]: result = function(*args, **kwargs) [ 767.042147] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 767.042147] env[61868]: return func(*args, **kwargs) [ 767.042147] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 767.042147] env[61868]: raise e [ 767.042147] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 767.042147] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 767.042147] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 767.042147] env[61868]: created_port_ids = self._update_ports_for_instance( [ 767.042147] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 767.042147] env[61868]: with excutils.save_and_reraise_exception(): [ 767.042147] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 767.042147] env[61868]: self.force_reraise() [ 767.042147] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 767.042147] env[61868]: raise self.value [ 767.042147] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 767.042147] env[61868]: updated_port = self._update_port( [ 767.042147] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 767.042147] env[61868]: _ensure_no_port_binding_failure(port) [ 767.042147] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 767.042147] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 767.043420] env[61868]: nova.exception.PortBindingFailed: Binding failed for port a9b0a8d5-d2f4-4b81-952f-f194085c3b38, please check neutron logs for more information. [ 767.043420] env[61868]: Removing descriptor: 15 [ 767.126133] env[61868]: DEBUG oslo_vmware.api [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': task-1315326, 'name': PowerOnVM_Task, 'duration_secs': 0.413113} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.126562] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 767.126883] env[61868]: INFO nova.compute.manager [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Took 4.82 seconds to spawn the instance on the hypervisor. [ 767.127213] env[61868]: DEBUG nova.compute.manager [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 767.128423] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-835040a9-257b-48f1-bf90-4d9497be5c18 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.132442] env[61868]: DEBUG nova.network.neutron [-] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.236582] env[61868]: ERROR nova.scheduler.client.report [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [req-0eb8df4d-b668-4474-a290-d59402df2518] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 83e67721-2ac9-4a23-aa31-82aca86979c8. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-0eb8df4d-b668-4474-a290-d59402df2518"}]} [ 767.254182] env[61868]: DEBUG nova.scheduler.client.report [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Refreshing inventories for resource provider 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 767.264193] env[61868]: DEBUG nova.compute.manager [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 767.268722] env[61868]: DEBUG nova.scheduler.client.report [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Updating ProviderTree inventory for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 767.268925] env[61868]: DEBUG nova.compute.provider_tree [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Updating inventory in ProviderTree for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 767.279421] env[61868]: DEBUG nova.scheduler.client.report [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Refreshing aggregate associations for resource provider 83e67721-2ac9-4a23-aa31-82aca86979c8, aggregates: None {{(pid=61868) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 767.289667] env[61868]: DEBUG nova.virt.hardware [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 767.289843] env[61868]: DEBUG nova.virt.hardware [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 767.289843] env[61868]: DEBUG nova.virt.hardware [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 767.290097] env[61868]: DEBUG nova.virt.hardware [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 767.290189] env[61868]: DEBUG nova.virt.hardware [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 767.290330] env[61868]: DEBUG nova.virt.hardware [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 767.290560] env[61868]: DEBUG nova.virt.hardware [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 767.290674] env[61868]: DEBUG nova.virt.hardware [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 767.290832] env[61868]: DEBUG nova.virt.hardware [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 767.290989] env[61868]: DEBUG nova.virt.hardware [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 767.291558] env[61868]: DEBUG nova.virt.hardware [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 767.292206] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-103b0d5e-215c-4264-ab54-7510314c6ae5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.299816] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2ba9248-01cc-435a-90a2-b3f9e172bd83 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.304183] env[61868]: DEBUG nova.scheduler.client.report [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Refreshing trait associations for resource provider 83e67721-2ac9-4a23-aa31-82aca86979c8, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61868) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 767.316211] env[61868]: ERROR nova.compute.manager [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a9b0a8d5-d2f4-4b81-952f-f194085c3b38, please check neutron logs for more information. [ 767.316211] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Traceback (most recent call last): [ 767.316211] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 767.316211] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] yield resources [ 767.316211] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 767.316211] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] self.driver.spawn(context, instance, image_meta, [ 767.316211] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 767.316211] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] self._vmops.spawn(context, instance, image_meta, injected_files, [ 767.316211] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 767.316211] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] vm_ref = self.build_virtual_machine(instance, [ 767.316211] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 767.316615] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] vif_infos = vmwarevif.get_vif_info(self._session, [ 767.316615] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 767.316615] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] for vif in network_info: [ 767.316615] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 767.316615] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] return self._sync_wrapper(fn, *args, **kwargs) [ 767.316615] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 767.316615] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] self.wait() [ 767.316615] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 767.316615] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] self[:] = self._gt.wait() [ 767.316615] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 767.316615] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] return self._exit_event.wait() [ 767.316615] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 767.316615] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] current.throw(*self._exc) [ 767.317066] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 767.317066] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] result = function(*args, **kwargs) [ 767.317066] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 767.317066] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] return func(*args, **kwargs) [ 767.317066] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 767.317066] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] raise e [ 767.317066] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 767.317066] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] nwinfo = self.network_api.allocate_for_instance( [ 767.317066] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 767.317066] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] created_port_ids = self._update_ports_for_instance( [ 767.317066] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 767.317066] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] with excutils.save_and_reraise_exception(): [ 767.317066] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 767.317444] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] self.force_reraise() [ 767.317444] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 767.317444] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] raise self.value [ 767.317444] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 767.317444] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] updated_port = self._update_port( [ 767.317444] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 767.317444] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] _ensure_no_port_binding_failure(port) [ 767.317444] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 767.317444] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] raise exception.PortBindingFailed(port_id=port['id']) [ 767.317444] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] nova.exception.PortBindingFailed: Binding failed for port a9b0a8d5-d2f4-4b81-952f-f194085c3b38, please check neutron logs for more information. [ 767.317444] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] [ 767.317444] env[61868]: INFO nova.compute.manager [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Terminating instance [ 767.318902] env[61868]: DEBUG oslo_concurrency.lockutils [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "refresh_cache-ebb09f50-50b8-49bb-b366-5b8906605699" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.318902] env[61868]: DEBUG oslo_concurrency.lockutils [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquired lock "refresh_cache-ebb09f50-50b8-49bb-b366-5b8906605699" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.318902] env[61868]: DEBUG nova.network.neutron [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 767.580418] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afcbb5ad-a11d-4e5a-80d1-0f9015a13bce {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.589025] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1338f512-ad61-44ea-b31d-4b67ba4c0bd4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.618614] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85bbd646-994b-4a88-88cd-cbcca6288690 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.625673] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b5295f3-a917-49f0-b93b-86d5e4ec3ec8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.638703] env[61868]: DEBUG nova.compute.provider_tree [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Updating inventory in ProviderTree for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 767.642693] env[61868]: INFO nova.compute.manager [-] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Took 1.04 seconds to deallocate network for instance. [ 767.648051] env[61868]: DEBUG nova.compute.claims [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 767.648232] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.650826] env[61868]: INFO nova.compute.manager [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Took 24.67 seconds to build instance. [ 767.838160] env[61868]: DEBUG nova.network.neutron [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 767.942540] env[61868]: DEBUG nova.network.neutron [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.152466] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b46c7d43-3f79-4072-a6df-3744331dd2a0 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Lock "4de32ee6-e205-435b-a75e-94c3121048e3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 137.692s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.177827] env[61868]: DEBUG nova.scheduler.client.report [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Updated inventory for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 with generation 77 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 768.178110] env[61868]: DEBUG nova.compute.provider_tree [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Updating resource provider 83e67721-2ac9-4a23-aa31-82aca86979c8 generation from 77 to 78 during operation: update_inventory {{(pid=61868) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 768.178299] env[61868]: DEBUG nova.compute.provider_tree [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Updating inventory in ProviderTree for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 768.396884] env[61868]: DEBUG nova.compute.manager [req-d6be67fd-c8a9-4c8d-b68a-65dfce551dcd req-ad0e7b38-0c1a-4128-8dab-266303df9272 service nova] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Received event network-changed-a9b0a8d5-d2f4-4b81-952f-f194085c3b38 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 768.396884] env[61868]: DEBUG nova.compute.manager [req-d6be67fd-c8a9-4c8d-b68a-65dfce551dcd req-ad0e7b38-0c1a-4128-8dab-266303df9272 service nova] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Refreshing instance network info cache due to event network-changed-a9b0a8d5-d2f4-4b81-952f-f194085c3b38. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 768.396884] env[61868]: DEBUG oslo_concurrency.lockutils [req-d6be67fd-c8a9-4c8d-b68a-65dfce551dcd req-ad0e7b38-0c1a-4128-8dab-266303df9272 service nova] Acquiring lock "refresh_cache-ebb09f50-50b8-49bb-b366-5b8906605699" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.445845] env[61868]: DEBUG oslo_concurrency.lockutils [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Releasing lock "refresh_cache-ebb09f50-50b8-49bb-b366-5b8906605699" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.446279] env[61868]: DEBUG nova.compute.manager [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 768.446496] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 768.446838] env[61868]: DEBUG oslo_concurrency.lockutils [req-d6be67fd-c8a9-4c8d-b68a-65dfce551dcd req-ad0e7b38-0c1a-4128-8dab-266303df9272 service nova] Acquired lock "refresh_cache-ebb09f50-50b8-49bb-b366-5b8906605699" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.447013] env[61868]: DEBUG nova.network.neutron [req-d6be67fd-c8a9-4c8d-b68a-65dfce551dcd req-ad0e7b38-0c1a-4128-8dab-266303df9272 service nova] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Refreshing network info cache for port a9b0a8d5-d2f4-4b81-952f-f194085c3b38 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 768.448057] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-380fe487-1c80-41cc-af32-864fdc185d12 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.457434] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f81cd38-3793-48c9-a77b-233cd0a27521 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.479898] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ebb09f50-50b8-49bb-b366-5b8906605699 could not be found. [ 768.480101] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 768.480280] env[61868]: INFO nova.compute.manager [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Took 0.03 seconds to destroy the instance on the hypervisor. [ 768.480505] env[61868]: DEBUG oslo.service.loopingcall [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 768.480707] env[61868]: DEBUG nova.compute.manager [-] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 768.480796] env[61868]: DEBUG nova.network.neutron [-] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 768.495203] env[61868]: DEBUG nova.network.neutron [-] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 768.634446] env[61868]: INFO nova.compute.manager [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Rebuilding instance [ 768.655352] env[61868]: DEBUG nova.compute.manager [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 768.678720] env[61868]: DEBUG nova.compute.manager [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 768.679609] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90b4bb9e-fd8f-47fa-b97e-e7c7c78831d1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.682911] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.443s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.683371] env[61868]: DEBUG nova.compute.manager [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 768.687028] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.499s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.968275] env[61868]: DEBUG nova.network.neutron [req-d6be67fd-c8a9-4c8d-b68a-65dfce551dcd req-ad0e7b38-0c1a-4128-8dab-266303df9272 service nova] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 768.997236] env[61868]: DEBUG nova.network.neutron [-] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.055337] env[61868]: DEBUG nova.network.neutron [req-d6be67fd-c8a9-4c8d-b68a-65dfce551dcd req-ad0e7b38-0c1a-4128-8dab-266303df9272 service nova] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.175876] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.191437] env[61868]: DEBUG nova.compute.utils [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 769.195664] env[61868]: DEBUG nova.compute.manager [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 769.195874] env[61868]: DEBUG nova.network.neutron [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 769.198601] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 769.198866] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-805e7d12-290a-455d-8581-3731388c17f7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.206828] env[61868]: DEBUG oslo_vmware.api [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Waiting for the task: (returnval){ [ 769.206828] env[61868]: value = "task-1315327" [ 769.206828] env[61868]: _type = "Task" [ 769.206828] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.214826] env[61868]: DEBUG oslo_vmware.api [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': task-1315327, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.242436] env[61868]: DEBUG nova.policy [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7f2eec1c3b204775ae5ffb09518c79e5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '50658d903dee454eb544ebf92621faa2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 769.500138] env[61868]: INFO nova.compute.manager [-] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Took 1.02 seconds to deallocate network for instance. [ 769.506895] env[61868]: DEBUG nova.compute.claims [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 769.507291] env[61868]: DEBUG oslo_concurrency.lockutils [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.508319] env[61868]: DEBUG nova.network.neutron [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Successfully created port: 7eec1a36-0cee-40ed-827c-749d6c1dab06 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 769.547059] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99c96756-d3dd-4ddb-9dee-d84f0d76a82f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.554813] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a390f0ca-3f00-4016-b678-25f9d73b94aa {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.558347] env[61868]: DEBUG oslo_concurrency.lockutils [req-d6be67fd-c8a9-4c8d-b68a-65dfce551dcd req-ad0e7b38-0c1a-4128-8dab-266303df9272 service nova] Releasing lock "refresh_cache-ebb09f50-50b8-49bb-b366-5b8906605699" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.558577] env[61868]: DEBUG nova.compute.manager [req-d6be67fd-c8a9-4c8d-b68a-65dfce551dcd req-ad0e7b38-0c1a-4128-8dab-266303df9272 service nova] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Received event network-vif-deleted-a9b0a8d5-d2f4-4b81-952f-f194085c3b38 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 769.587320] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1928929c-db9d-4f3c-b787-d3ee4c1dd235 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.594781] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-887dc414-6a71-40fe-aa78-b06c2fcbce22 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.607871] env[61868]: DEBUG nova.compute.provider_tree [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 769.697027] env[61868]: DEBUG nova.compute.manager [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 769.720156] env[61868]: DEBUG oslo_vmware.api [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': task-1315327, 'name': PowerOffVM_Task, 'duration_secs': 0.109381} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.720505] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 769.721048] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 769.722774] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15e5c811-a9ce-4d03-acab-687b5435f18b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.729370] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 769.729606] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c7d22421-af0f-4202-aabb-dc4397506ccd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.754168] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 769.754403] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 769.754580] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Deleting the datastore file [datastore1] 4de32ee6-e205-435b-a75e-94c3121048e3 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 769.754822] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-de51b806-3509-46ae-b2cf-aebcec4f8969 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.761334] env[61868]: DEBUG oslo_vmware.api [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Waiting for the task: (returnval){ [ 769.761334] env[61868]: value = "task-1315329" [ 769.761334] env[61868]: _type = "Task" [ 769.761334] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.769877] env[61868]: DEBUG oslo_vmware.api [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': task-1315329, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.111429] env[61868]: DEBUG nova.scheduler.client.report [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 770.273344] env[61868]: DEBUG oslo_vmware.api [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': task-1315329, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.093571} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.273344] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 770.273344] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 770.273344] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 770.373385] env[61868]: ERROR nova.compute.manager [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7eec1a36-0cee-40ed-827c-749d6c1dab06, please check neutron logs for more information. [ 770.373385] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 770.373385] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 770.373385] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 770.373385] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 770.373385] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 770.373385] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 770.373385] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 770.373385] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 770.373385] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 770.373385] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 770.373385] env[61868]: ERROR nova.compute.manager raise self.value [ 770.373385] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 770.373385] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 770.373385] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 770.373385] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 770.373912] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 770.373912] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 770.373912] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7eec1a36-0cee-40ed-827c-749d6c1dab06, please check neutron logs for more information. [ 770.373912] env[61868]: ERROR nova.compute.manager [ 770.373912] env[61868]: Traceback (most recent call last): [ 770.373912] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 770.373912] env[61868]: listener.cb(fileno) [ 770.373912] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 770.373912] env[61868]: result = function(*args, **kwargs) [ 770.373912] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 770.373912] env[61868]: return func(*args, **kwargs) [ 770.373912] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 770.373912] env[61868]: raise e [ 770.373912] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 770.373912] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 770.373912] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 770.373912] env[61868]: created_port_ids = self._update_ports_for_instance( [ 770.373912] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 770.373912] env[61868]: with excutils.save_and_reraise_exception(): [ 770.373912] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 770.373912] env[61868]: self.force_reraise() [ 770.373912] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 770.373912] env[61868]: raise self.value [ 770.373912] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 770.373912] env[61868]: updated_port = self._update_port( [ 770.373912] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 770.373912] env[61868]: _ensure_no_port_binding_failure(port) [ 770.373912] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 770.373912] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 770.374830] env[61868]: nova.exception.PortBindingFailed: Binding failed for port 7eec1a36-0cee-40ed-827c-749d6c1dab06, please check neutron logs for more information. [ 770.374830] env[61868]: Removing descriptor: 15 [ 770.467247] env[61868]: DEBUG nova.compute.manager [req-6ab6d486-d0b9-4c8b-b208-a8040eada142 req-af9008f7-6986-4e51-ac6c-079eaa09c561 service nova] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Received event network-changed-7eec1a36-0cee-40ed-827c-749d6c1dab06 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 770.467463] env[61868]: DEBUG nova.compute.manager [req-6ab6d486-d0b9-4c8b-b208-a8040eada142 req-af9008f7-6986-4e51-ac6c-079eaa09c561 service nova] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Refreshing instance network info cache due to event network-changed-7eec1a36-0cee-40ed-827c-749d6c1dab06. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 770.467676] env[61868]: DEBUG oslo_concurrency.lockutils [req-6ab6d486-d0b9-4c8b-b208-a8040eada142 req-af9008f7-6986-4e51-ac6c-079eaa09c561 service nova] Acquiring lock "refresh_cache-29530178-d69c-4aed-9061-d3d1cfa954d0" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.468150] env[61868]: DEBUG oslo_concurrency.lockutils [req-6ab6d486-d0b9-4c8b-b208-a8040eada142 req-af9008f7-6986-4e51-ac6c-079eaa09c561 service nova] Acquired lock "refresh_cache-29530178-d69c-4aed-9061-d3d1cfa954d0" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.468150] env[61868]: DEBUG nova.network.neutron [req-6ab6d486-d0b9-4c8b-b208-a8040eada142 req-af9008f7-6986-4e51-ac6c-079eaa09c561 service nova] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Refreshing network info cache for port 7eec1a36-0cee-40ed-827c-749d6c1dab06 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 770.616846] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.930s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.617546] env[61868]: ERROR nova.compute.manager [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 898e66dc-6927-44a5-9a0a-1a0114d58d14, please check neutron logs for more information. [ 770.617546] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Traceback (most recent call last): [ 770.617546] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 770.617546] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] self.driver.spawn(context, instance, image_meta, [ 770.617546] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 770.617546] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] self._vmops.spawn(context, instance, image_meta, injected_files, [ 770.617546] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 770.617546] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] vm_ref = self.build_virtual_machine(instance, [ 770.617546] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 770.617546] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] vif_infos = vmwarevif.get_vif_info(self._session, [ 770.617546] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 770.617894] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] for vif in network_info: [ 770.617894] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 770.617894] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] return self._sync_wrapper(fn, *args, **kwargs) [ 770.617894] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 770.617894] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] self.wait() [ 770.617894] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 770.617894] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] self[:] = self._gt.wait() [ 770.617894] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 770.617894] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] return self._exit_event.wait() [ 770.617894] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 770.617894] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] result = hub.switch() [ 770.617894] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 770.617894] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] return self.greenlet.switch() [ 770.618284] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 770.618284] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] result = function(*args, **kwargs) [ 770.618284] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 770.618284] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] return func(*args, **kwargs) [ 770.618284] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 770.618284] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] raise e [ 770.618284] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 770.618284] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] nwinfo = self.network_api.allocate_for_instance( [ 770.618284] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 770.618284] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] created_port_ids = self._update_ports_for_instance( [ 770.618284] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 770.618284] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] with excutils.save_and_reraise_exception(): [ 770.618284] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 770.618709] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] self.force_reraise() [ 770.618709] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 770.618709] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] raise self.value [ 770.618709] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 770.618709] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] updated_port = self._update_port( [ 770.618709] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 770.618709] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] _ensure_no_port_binding_failure(port) [ 770.618709] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 770.618709] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] raise exception.PortBindingFailed(port_id=port['id']) [ 770.618709] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] nova.exception.PortBindingFailed: Binding failed for port 898e66dc-6927-44a5-9a0a-1a0114d58d14, please check neutron logs for more information. [ 770.618709] env[61868]: ERROR nova.compute.manager [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] [ 770.619063] env[61868]: DEBUG nova.compute.utils [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Binding failed for port 898e66dc-6927-44a5-9a0a-1a0114d58d14, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 770.619850] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.207s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.621426] env[61868]: INFO nova.compute.claims [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 770.624430] env[61868]: DEBUG nova.compute.manager [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Build of instance 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea was re-scheduled: Binding failed for port 898e66dc-6927-44a5-9a0a-1a0114d58d14, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 770.624565] env[61868]: DEBUG nova.compute.manager [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 770.624789] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Acquiring lock "refresh_cache-06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.624966] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Acquired lock "refresh_cache-06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.625207] env[61868]: DEBUG nova.network.neutron [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 770.706571] env[61868]: DEBUG nova.compute.manager [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 770.764862] env[61868]: DEBUG nova.virt.hardware [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 770.765201] env[61868]: DEBUG nova.virt.hardware [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 770.765396] env[61868]: DEBUG nova.virt.hardware [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 770.765606] env[61868]: DEBUG nova.virt.hardware [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 770.765776] env[61868]: DEBUG nova.virt.hardware [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 770.765938] env[61868]: DEBUG nova.virt.hardware [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 770.766189] env[61868]: DEBUG nova.virt.hardware [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 770.766387] env[61868]: DEBUG nova.virt.hardware [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 770.766557] env[61868]: DEBUG nova.virt.hardware [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 770.766804] env[61868]: DEBUG nova.virt.hardware [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 770.766963] env[61868]: DEBUG nova.virt.hardware [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 770.767947] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac636879-6163-4443-9435-bc95dfc23b89 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.776899] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c416ca5-4dd7-474a-8eed-f9f039b1d3cc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.796029] env[61868]: ERROR nova.compute.manager [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7eec1a36-0cee-40ed-827c-749d6c1dab06, please check neutron logs for more information. [ 770.796029] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Traceback (most recent call last): [ 770.796029] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 770.796029] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] yield resources [ 770.796029] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 770.796029] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] self.driver.spawn(context, instance, image_meta, [ 770.796029] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 770.796029] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 770.796029] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 770.796029] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] vm_ref = self.build_virtual_machine(instance, [ 770.796029] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 770.797072] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] vif_infos = vmwarevif.get_vif_info(self._session, [ 770.797072] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 770.797072] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] for vif in network_info: [ 770.797072] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 770.797072] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] return self._sync_wrapper(fn, *args, **kwargs) [ 770.797072] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 770.797072] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] self.wait() [ 770.797072] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 770.797072] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] self[:] = self._gt.wait() [ 770.797072] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 770.797072] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] return self._exit_event.wait() [ 770.797072] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 770.797072] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] current.throw(*self._exc) [ 770.797459] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 770.797459] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] result = function(*args, **kwargs) [ 770.797459] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 770.797459] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] return func(*args, **kwargs) [ 770.797459] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 770.797459] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] raise e [ 770.797459] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 770.797459] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] nwinfo = self.network_api.allocate_for_instance( [ 770.797459] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 770.797459] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] created_port_ids = self._update_ports_for_instance( [ 770.797459] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 770.797459] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] with excutils.save_and_reraise_exception(): [ 770.797459] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 770.798046] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] self.force_reraise() [ 770.798046] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 770.798046] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] raise self.value [ 770.798046] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 770.798046] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] updated_port = self._update_port( [ 770.798046] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 770.798046] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] _ensure_no_port_binding_failure(port) [ 770.798046] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 770.798046] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] raise exception.PortBindingFailed(port_id=port['id']) [ 770.798046] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] nova.exception.PortBindingFailed: Binding failed for port 7eec1a36-0cee-40ed-827c-749d6c1dab06, please check neutron logs for more information. [ 770.798046] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] [ 770.798046] env[61868]: INFO nova.compute.manager [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Terminating instance [ 770.798666] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Acquiring lock "refresh_cache-29530178-d69c-4aed-9061-d3d1cfa954d0" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.988330] env[61868]: DEBUG nova.network.neutron [req-6ab6d486-d0b9-4c8b-b208-a8040eada142 req-af9008f7-6986-4e51-ac6c-079eaa09c561 service nova] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 771.063044] env[61868]: DEBUG nova.network.neutron [req-6ab6d486-d0b9-4c8b-b208-a8040eada142 req-af9008f7-6986-4e51-ac6c-079eaa09c561 service nova] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.146055] env[61868]: DEBUG nova.network.neutron [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 771.215758] env[61868]: DEBUG nova.network.neutron [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.306510] env[61868]: DEBUG nova.virt.hardware [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 771.306843] env[61868]: DEBUG nova.virt.hardware [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 771.306892] env[61868]: DEBUG nova.virt.hardware [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 771.307079] env[61868]: DEBUG nova.virt.hardware [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 771.307226] env[61868]: DEBUG nova.virt.hardware [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 771.307369] env[61868]: DEBUG nova.virt.hardware [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 771.307573] env[61868]: DEBUG nova.virt.hardware [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 771.307726] env[61868]: DEBUG nova.virt.hardware [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 771.307886] env[61868]: DEBUG nova.virt.hardware [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 771.308055] env[61868]: DEBUG nova.virt.hardware [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 771.308240] env[61868]: DEBUG nova.virt.hardware [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 771.309280] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-735a6e3c-c6c2-4889-8ea0-beab3d44f9a4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.317147] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eda65a86-cbe5-4569-859c-0277d135fdb7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.330213] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Instance VIF info [] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 771.335907] env[61868]: DEBUG oslo.service.loopingcall [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 771.336138] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 771.336346] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-72bf878a-7ebe-44ed-be64-ab12e9685ea8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.352472] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 771.352472] env[61868]: value = "task-1315330" [ 771.352472] env[61868]: _type = "Task" [ 771.352472] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.359931] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315330, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.566672] env[61868]: DEBUG oslo_concurrency.lockutils [req-6ab6d486-d0b9-4c8b-b208-a8040eada142 req-af9008f7-6986-4e51-ac6c-079eaa09c561 service nova] Releasing lock "refresh_cache-29530178-d69c-4aed-9061-d3d1cfa954d0" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.566957] env[61868]: DEBUG nova.compute.manager [req-6ab6d486-d0b9-4c8b-b208-a8040eada142 req-af9008f7-6986-4e51-ac6c-079eaa09c561 service nova] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Received event network-vif-deleted-7eec1a36-0cee-40ed-827c-749d6c1dab06 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 771.567394] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Acquired lock "refresh_cache-29530178-d69c-4aed-9061-d3d1cfa954d0" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.567589] env[61868]: DEBUG nova.network.neutron [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 771.719016] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Releasing lock "refresh_cache-06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.719383] env[61868]: DEBUG nova.compute.manager [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 771.719466] env[61868]: DEBUG nova.compute.manager [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 771.719663] env[61868]: DEBUG nova.network.neutron [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 771.735759] env[61868]: DEBUG nova.network.neutron [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 771.863050] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315330, 'name': CreateVM_Task, 'duration_secs': 0.29014} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.865183] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 771.865755] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.865991] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.866324] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 771.866582] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-243648d9-ec8e-4d26-9dbe-0a42c54dad15 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.871015] env[61868]: DEBUG oslo_vmware.api [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Waiting for the task: (returnval){ [ 771.871015] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52deb13c-6c7e-11f6-b8c7-0b2a2076785e" [ 771.871015] env[61868]: _type = "Task" [ 771.871015] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.880648] env[61868]: DEBUG oslo_vmware.api [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52deb13c-6c7e-11f6-b8c7-0b2a2076785e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.944071] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5af40b7e-eb24-4cd0-b901-609ca73ee9c8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.954296] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85cf6dd5-4d69-41ab-a396-6494ce8feff1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.994333] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1142796a-f282-4a0e-8a4e-2ef75c9470d1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.001700] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7b7aa83-b227-4f8d-9130-a7e9faac275a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.015029] env[61868]: DEBUG nova.compute.provider_tree [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 772.088054] env[61868]: DEBUG nova.network.neutron [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 772.167486] env[61868]: DEBUG nova.network.neutron [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.238032] env[61868]: DEBUG nova.network.neutron [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.381378] env[61868]: DEBUG oslo_vmware.api [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52deb13c-6c7e-11f6-b8c7-0b2a2076785e, 'name': SearchDatastore_Task, 'duration_secs': 0.012376} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.381678] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.381902] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 772.382169] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.382318] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.382495] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 772.382737] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b3303ceb-1cbf-4520-bb71-9c4bb73022b9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.391815] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 772.391983] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 772.392659] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41492b3f-da9e-4d00-9263-9fd09efd4673 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.397111] env[61868]: DEBUG oslo_vmware.api [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Waiting for the task: (returnval){ [ 772.397111] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]521d41f9-9b9c-3784-b5c0-73e8f82ab9d7" [ 772.397111] env[61868]: _type = "Task" [ 772.397111] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.403978] env[61868]: DEBUG oslo_vmware.api [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]521d41f9-9b9c-3784-b5c0-73e8f82ab9d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.518125] env[61868]: DEBUG nova.scheduler.client.report [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 772.670622] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Releasing lock "refresh_cache-29530178-d69c-4aed-9061-d3d1cfa954d0" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.671119] env[61868]: DEBUG nova.compute.manager [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 772.671335] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 772.671625] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-95a6637a-bc8d-4ad7-9dad-d14b0b69d0b8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.680515] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-379dfdd4-7da7-451e-ad5d-483694900ef6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.701397] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 29530178-d69c-4aed-9061-d3d1cfa954d0 could not be found. [ 772.701606] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 772.701783] env[61868]: INFO nova.compute.manager [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Took 0.03 seconds to destroy the instance on the hypervisor. [ 772.702058] env[61868]: DEBUG oslo.service.loopingcall [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 772.702285] env[61868]: DEBUG nova.compute.manager [-] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 772.702385] env[61868]: DEBUG nova.network.neutron [-] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 772.716422] env[61868]: DEBUG nova.network.neutron [-] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 772.740151] env[61868]: INFO nova.compute.manager [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea] Took 1.02 seconds to deallocate network for instance. [ 772.908044] env[61868]: DEBUG oslo_vmware.api [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]521d41f9-9b9c-3784-b5c0-73e8f82ab9d7, 'name': SearchDatastore_Task, 'duration_secs': 0.008319} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.908292] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a5ab1249-cf51-4fd0-bcd9-c3b15fd7903d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.913443] env[61868]: DEBUG oslo_vmware.api [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Waiting for the task: (returnval){ [ 772.913443] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]525dbcc5-2d65-cb40-d00f-2d3501d95978" [ 772.913443] env[61868]: _type = "Task" [ 772.913443] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.920887] env[61868]: DEBUG oslo_vmware.api [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]525dbcc5-2d65-cb40-d00f-2d3501d95978, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.024024] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.404s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.024488] env[61868]: DEBUG nova.compute.manager [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 773.027129] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.477s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.219786] env[61868]: DEBUG nova.network.neutron [-] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.424128] env[61868]: DEBUG oslo_vmware.api [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]525dbcc5-2d65-cb40-d00f-2d3501d95978, 'name': SearchDatastore_Task, 'duration_secs': 0.008484} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.424433] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.424690] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 4de32ee6-e205-435b-a75e-94c3121048e3/4de32ee6-e205-435b-a75e-94c3121048e3.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 773.424935] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-45abe265-9414-49e0-8f48-8c22e2db30bc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.430577] env[61868]: DEBUG oslo_vmware.api [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Waiting for the task: (returnval){ [ 773.430577] env[61868]: value = "task-1315331" [ 773.430577] env[61868]: _type = "Task" [ 773.430577] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.437745] env[61868]: DEBUG oslo_vmware.api [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': task-1315331, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.532792] env[61868]: DEBUG nova.compute.utils [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 773.537399] env[61868]: DEBUG nova.compute.manager [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 773.537399] env[61868]: DEBUG nova.network.neutron [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 773.581468] env[61868]: DEBUG nova.policy [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f1a0761d7c814c008b776c5c4f91f02b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5f78b988969b414f9461671f8a8ff225', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 773.724118] env[61868]: INFO nova.compute.manager [-] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Took 1.02 seconds to deallocate network for instance. [ 773.731756] env[61868]: DEBUG nova.compute.claims [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 773.731975] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 773.777335] env[61868]: INFO nova.scheduler.client.report [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Deleted allocations for instance 06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea [ 773.917411] env[61868]: DEBUG nova.network.neutron [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Successfully created port: 651014ed-1297-4d85-967c-03c06efe3fc2 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 773.943757] env[61868]: DEBUG oslo_vmware.api [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': task-1315331, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.480043} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.944092] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 4de32ee6-e205-435b-a75e-94c3121048e3/4de32ee6-e205-435b-a75e-94c3121048e3.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 773.944347] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 773.944650] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ac4bbba8-03df-42b8-b747-3aadbb9f0fe0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.950836] env[61868]: DEBUG oslo_vmware.api [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Waiting for the task: (returnval){ [ 773.950836] env[61868]: value = "task-1315332" [ 773.950836] env[61868]: _type = "Task" [ 773.950836] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.952442] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f86f08e8-9bed-4a3a-b1c8-9420940c5112 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.966026] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-720ccea4-e477-47ea-b82d-fb45c0dd7e71 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.968997] env[61868]: DEBUG oslo_vmware.api [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': task-1315332, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.997595] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbf1bc5e-aaa7-4671-95f1-a3c83ef69304 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.004787] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ae49ed0-fb01-48e3-a568-1ea819de7f76 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.018467] env[61868]: DEBUG nova.compute.provider_tree [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 774.037046] env[61868]: DEBUG nova.compute.manager [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 774.287677] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4a25f413-9a9c-4560-9f81-bf973fd484bc tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Lock "06336a4e-b9e8-4036-8b4d-6e9a2ccc6bea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 150.690s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.461518] env[61868]: DEBUG oslo_vmware.api [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': task-1315332, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063443} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.461842] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 774.462706] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c5efee7-f5de-4494-b684-4a3ac14a7249 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.482389] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Reconfiguring VM instance instance-0000002c to attach disk [datastore1] 4de32ee6-e205-435b-a75e-94c3121048e3/4de32ee6-e205-435b-a75e-94c3121048e3.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 774.482686] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1797a88d-3c7a-4a3a-b5b8-7e935cf9b50b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.501412] env[61868]: DEBUG oslo_vmware.api [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Waiting for the task: (returnval){ [ 774.501412] env[61868]: value = "task-1315333" [ 774.501412] env[61868]: _type = "Task" [ 774.501412] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.509026] env[61868]: DEBUG oslo_vmware.api [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': task-1315333, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.522921] env[61868]: DEBUG nova.scheduler.client.report [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 774.714322] env[61868]: DEBUG nova.compute.manager [req-35616e47-5b4f-4d50-a58c-26ee65dc9135 req-53d1673c-0318-4fdf-a6a0-e15f96e0da30 service nova] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Received event network-changed-651014ed-1297-4d85-967c-03c06efe3fc2 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 774.714525] env[61868]: DEBUG nova.compute.manager [req-35616e47-5b4f-4d50-a58c-26ee65dc9135 req-53d1673c-0318-4fdf-a6a0-e15f96e0da30 service nova] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Refreshing instance network info cache due to event network-changed-651014ed-1297-4d85-967c-03c06efe3fc2. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 774.714735] env[61868]: DEBUG oslo_concurrency.lockutils [req-35616e47-5b4f-4d50-a58c-26ee65dc9135 req-53d1673c-0318-4fdf-a6a0-e15f96e0da30 service nova] Acquiring lock "refresh_cache-e023cfb8-5f1c-41d9-9e43-d10e073234ea" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.714876] env[61868]: DEBUG oslo_concurrency.lockutils [req-35616e47-5b4f-4d50-a58c-26ee65dc9135 req-53d1673c-0318-4fdf-a6a0-e15f96e0da30 service nova] Acquired lock "refresh_cache-e023cfb8-5f1c-41d9-9e43-d10e073234ea" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.715106] env[61868]: DEBUG nova.network.neutron [req-35616e47-5b4f-4d50-a58c-26ee65dc9135 req-53d1673c-0318-4fdf-a6a0-e15f96e0da30 service nova] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Refreshing network info cache for port 651014ed-1297-4d85-967c-03c06efe3fc2 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 774.792618] env[61868]: DEBUG nova.compute.manager [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 774.899517] env[61868]: ERROR nova.compute.manager [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 651014ed-1297-4d85-967c-03c06efe3fc2, please check neutron logs for more information. [ 774.899517] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 774.899517] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 774.899517] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 774.899517] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 774.899517] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 774.899517] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 774.899517] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 774.899517] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 774.899517] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 774.899517] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 774.899517] env[61868]: ERROR nova.compute.manager raise self.value [ 774.899517] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 774.899517] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 774.899517] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 774.899517] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 774.900140] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 774.900140] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 774.900140] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 651014ed-1297-4d85-967c-03c06efe3fc2, please check neutron logs for more information. [ 774.900140] env[61868]: ERROR nova.compute.manager [ 774.900140] env[61868]: Traceback (most recent call last): [ 774.900140] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 774.900140] env[61868]: listener.cb(fileno) [ 774.900140] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 774.900140] env[61868]: result = function(*args, **kwargs) [ 774.900140] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 774.900140] env[61868]: return func(*args, **kwargs) [ 774.900140] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 774.900140] env[61868]: raise e [ 774.900140] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 774.900140] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 774.900140] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 774.900140] env[61868]: created_port_ids = self._update_ports_for_instance( [ 774.900140] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 774.900140] env[61868]: with excutils.save_and_reraise_exception(): [ 774.900140] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 774.900140] env[61868]: self.force_reraise() [ 774.900140] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 774.900140] env[61868]: raise self.value [ 774.900140] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 774.900140] env[61868]: updated_port = self._update_port( [ 774.900140] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 774.900140] env[61868]: _ensure_no_port_binding_failure(port) [ 774.900140] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 774.900140] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 774.901155] env[61868]: nova.exception.PortBindingFailed: Binding failed for port 651014ed-1297-4d85-967c-03c06efe3fc2, please check neutron logs for more information. [ 774.901155] env[61868]: Removing descriptor: 15 [ 775.010910] env[61868]: DEBUG oslo_vmware.api [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': task-1315333, 'name': ReconfigVM_Task, 'duration_secs': 0.298816} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.011606] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Reconfigured VM instance instance-0000002c to attach disk [datastore1] 4de32ee6-e205-435b-a75e-94c3121048e3/4de32ee6-e205-435b-a75e-94c3121048e3.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 775.012098] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e85a8161-39d1-4e30-87f3-8e3974862b13 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.017575] env[61868]: DEBUG oslo_vmware.api [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Waiting for the task: (returnval){ [ 775.017575] env[61868]: value = "task-1315334" [ 775.017575] env[61868]: _type = "Task" [ 775.017575] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.024787] env[61868]: DEBUG oslo_vmware.api [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': task-1315334, 'name': Rename_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.026522] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.999s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.027099] env[61868]: ERROR nova.compute.manager [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7192ca4b-ab2e-4d7c-8dec-36372e362ecc, please check neutron logs for more information. [ 775.027099] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Traceback (most recent call last): [ 775.027099] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 775.027099] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] self.driver.spawn(context, instance, image_meta, [ 775.027099] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 775.027099] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 775.027099] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 775.027099] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] vm_ref = self.build_virtual_machine(instance, [ 775.027099] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 775.027099] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] vif_infos = vmwarevif.get_vif_info(self._session, [ 775.027099] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 775.027474] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] for vif in network_info: [ 775.027474] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 775.027474] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] return self._sync_wrapper(fn, *args, **kwargs) [ 775.027474] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 775.027474] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] self.wait() [ 775.027474] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 775.027474] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] self[:] = self._gt.wait() [ 775.027474] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 775.027474] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] return self._exit_event.wait() [ 775.027474] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 775.027474] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] result = hub.switch() [ 775.027474] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 775.027474] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] return self.greenlet.switch() [ 775.027896] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 775.027896] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] result = function(*args, **kwargs) [ 775.027896] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 775.027896] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] return func(*args, **kwargs) [ 775.027896] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 775.027896] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] raise e [ 775.027896] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 775.027896] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] nwinfo = self.network_api.allocate_for_instance( [ 775.027896] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 775.027896] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] created_port_ids = self._update_ports_for_instance( [ 775.027896] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 775.027896] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] with excutils.save_and_reraise_exception(): [ 775.027896] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 775.028334] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] self.force_reraise() [ 775.028334] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 775.028334] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] raise self.value [ 775.028334] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 775.028334] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] updated_port = self._update_port( [ 775.028334] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 775.028334] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] _ensure_no_port_binding_failure(port) [ 775.028334] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 775.028334] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] raise exception.PortBindingFailed(port_id=port['id']) [ 775.028334] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] nova.exception.PortBindingFailed: Binding failed for port 7192ca4b-ab2e-4d7c-8dec-36372e362ecc, please check neutron logs for more information. [ 775.028334] env[61868]: ERROR nova.compute.manager [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] [ 775.028695] env[61868]: DEBUG nova.compute.utils [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Binding failed for port 7192ca4b-ab2e-4d7c-8dec-36372e362ecc, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 775.028871] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.085s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.030737] env[61868]: INFO nova.compute.claims [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 775.033909] env[61868]: DEBUG nova.compute.manager [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Build of instance 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3 was re-scheduled: Binding failed for port 7192ca4b-ab2e-4d7c-8dec-36372e362ecc, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 775.034291] env[61868]: DEBUG nova.compute.manager [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 775.034523] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Acquiring lock "refresh_cache-2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.034674] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Acquired lock "refresh_cache-2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.034830] env[61868]: DEBUG nova.network.neutron [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 775.046110] env[61868]: DEBUG nova.compute.manager [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 775.071543] env[61868]: DEBUG nova.virt.hardware [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 775.071788] env[61868]: DEBUG nova.virt.hardware [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 775.071943] env[61868]: DEBUG nova.virt.hardware [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 775.072149] env[61868]: DEBUG nova.virt.hardware [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 775.072299] env[61868]: DEBUG nova.virt.hardware [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 775.072444] env[61868]: DEBUG nova.virt.hardware [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 775.072649] env[61868]: DEBUG nova.virt.hardware [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 775.072843] env[61868]: DEBUG nova.virt.hardware [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 775.073028] env[61868]: DEBUG nova.virt.hardware [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 775.073193] env[61868]: DEBUG nova.virt.hardware [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 775.073367] env[61868]: DEBUG nova.virt.hardware [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 775.074266] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2031665-592f-4d5e-a98e-25cc5cae4701 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.082416] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84648f42-e036-4d96-b0a3-999ca8bde2a9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.096272] env[61868]: ERROR nova.compute.manager [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 651014ed-1297-4d85-967c-03c06efe3fc2, please check neutron logs for more information. [ 775.096272] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Traceback (most recent call last): [ 775.096272] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 775.096272] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] yield resources [ 775.096272] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 775.096272] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] self.driver.spawn(context, instance, image_meta, [ 775.096272] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 775.096272] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] self._vmops.spawn(context, instance, image_meta, injected_files, [ 775.096272] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 775.096272] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] vm_ref = self.build_virtual_machine(instance, [ 775.096272] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 775.096674] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] vif_infos = vmwarevif.get_vif_info(self._session, [ 775.096674] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 775.096674] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] for vif in network_info: [ 775.096674] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 775.096674] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] return self._sync_wrapper(fn, *args, **kwargs) [ 775.096674] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 775.096674] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] self.wait() [ 775.096674] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 775.096674] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] self[:] = self._gt.wait() [ 775.096674] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 775.096674] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] return self._exit_event.wait() [ 775.096674] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 775.096674] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] current.throw(*self._exc) [ 775.097265] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 775.097265] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] result = function(*args, **kwargs) [ 775.097265] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 775.097265] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] return func(*args, **kwargs) [ 775.097265] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 775.097265] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] raise e [ 775.097265] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 775.097265] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] nwinfo = self.network_api.allocate_for_instance( [ 775.097265] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 775.097265] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] created_port_ids = self._update_ports_for_instance( [ 775.097265] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 775.097265] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] with excutils.save_and_reraise_exception(): [ 775.097265] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 775.097694] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] self.force_reraise() [ 775.097694] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 775.097694] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] raise self.value [ 775.097694] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 775.097694] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] updated_port = self._update_port( [ 775.097694] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 775.097694] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] _ensure_no_port_binding_failure(port) [ 775.097694] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 775.097694] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] raise exception.PortBindingFailed(port_id=port['id']) [ 775.097694] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] nova.exception.PortBindingFailed: Binding failed for port 651014ed-1297-4d85-967c-03c06efe3fc2, please check neutron logs for more information. [ 775.097694] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] [ 775.097694] env[61868]: INFO nova.compute.manager [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Terminating instance [ 775.098842] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] Acquiring lock "refresh_cache-e023cfb8-5f1c-41d9-9e43-d10e073234ea" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.239152] env[61868]: DEBUG nova.network.neutron [req-35616e47-5b4f-4d50-a58c-26ee65dc9135 req-53d1673c-0318-4fdf-a6a0-e15f96e0da30 service nova] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 775.312786] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.326558] env[61868]: DEBUG nova.network.neutron [req-35616e47-5b4f-4d50-a58c-26ee65dc9135 req-53d1673c-0318-4fdf-a6a0-e15f96e0da30 service nova] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.528261] env[61868]: DEBUG oslo_vmware.api [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': task-1315334, 'name': Rename_Task, 'duration_secs': 0.143701} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.528568] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 775.528756] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f864dd2c-a3a5-4a2c-8c1e-62fc09c1dbf2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.536526] env[61868]: DEBUG oslo_vmware.api [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Waiting for the task: (returnval){ [ 775.536526] env[61868]: value = "task-1315335" [ 775.536526] env[61868]: _type = "Task" [ 775.536526] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.546905] env[61868]: DEBUG oslo_vmware.api [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': task-1315335, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.555587] env[61868]: DEBUG nova.network.neutron [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 775.635133] env[61868]: DEBUG nova.network.neutron [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.830292] env[61868]: DEBUG oslo_concurrency.lockutils [req-35616e47-5b4f-4d50-a58c-26ee65dc9135 req-53d1673c-0318-4fdf-a6a0-e15f96e0da30 service nova] Releasing lock "refresh_cache-e023cfb8-5f1c-41d9-9e43-d10e073234ea" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.830817] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] Acquired lock "refresh_cache-e023cfb8-5f1c-41d9-9e43-d10e073234ea" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.831121] env[61868]: DEBUG nova.network.neutron [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 776.049276] env[61868]: DEBUG oslo_vmware.api [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': task-1315335, 'name': PowerOnVM_Task, 'duration_secs': 0.410427} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.049554] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 776.049797] env[61868]: DEBUG nova.compute.manager [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 776.050594] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b7cc23c-1981-4fb2-be83-6f6109eebf41 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.138026] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Releasing lock "refresh_cache-2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.138197] env[61868]: DEBUG nova.compute.manager [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 776.138380] env[61868]: DEBUG nova.compute.manager [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 776.138542] env[61868]: DEBUG nova.network.neutron [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 776.156275] env[61868]: DEBUG nova.network.neutron [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 776.348697] env[61868]: DEBUG nova.network.neutron [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 776.361515] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88fade57-4ec8-41bf-958d-d3c0a7aa364a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.368508] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc9852c9-b342-4869-979e-bb29036f392c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.400178] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-779403a0-87ba-49a2-8e62-3807c93f4bc1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.407228] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95c2bfcf-178a-4cb8-8333-9389818ad801 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.420156] env[61868]: DEBUG nova.compute.provider_tree [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 776.437310] env[61868]: DEBUG nova.network.neutron [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.569250] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.658331] env[61868]: DEBUG nova.network.neutron [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.762894] env[61868]: DEBUG nova.compute.manager [req-17876b8e-90a1-47c2-84ae-dbbe271c7ec1 req-e6980261-1ea7-4209-9ed7-fd6bd182753e service nova] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Received event network-vif-deleted-651014ed-1297-4d85-967c-03c06efe3fc2 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 776.923554] env[61868]: DEBUG nova.scheduler.client.report [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 776.940125] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] Releasing lock "refresh_cache-e023cfb8-5f1c-41d9-9e43-d10e073234ea" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.940544] env[61868]: DEBUG nova.compute.manager [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 776.940737] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 776.941717] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4678015c-f376-4e1a-ac81-6ded2fdfe67d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.950745] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f77a09d5-ff6e-47fb-b2b4-779ba1b7a4a1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.972979] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e023cfb8-5f1c-41d9-9e43-d10e073234ea could not be found. [ 776.973199] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 776.973376] env[61868]: INFO nova.compute.manager [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Took 0.03 seconds to destroy the instance on the hypervisor. [ 776.974410] env[61868]: DEBUG oslo.service.loopingcall [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 776.974410] env[61868]: DEBUG nova.compute.manager [-] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 776.974410] env[61868]: DEBUG nova.network.neutron [-] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 776.988332] env[61868]: DEBUG nova.network.neutron [-] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 777.160889] env[61868]: INFO nova.compute.manager [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] [instance: 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3] Took 1.02 seconds to deallocate network for instance. [ 777.291241] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3c4ba62e-0cd1-441f-bc9a-966fa2cf9b11 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Acquiring lock "4de32ee6-e205-435b-a75e-94c3121048e3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.291542] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3c4ba62e-0cd1-441f-bc9a-966fa2cf9b11 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Lock "4de32ee6-e205-435b-a75e-94c3121048e3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.291749] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3c4ba62e-0cd1-441f-bc9a-966fa2cf9b11 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Acquiring lock "4de32ee6-e205-435b-a75e-94c3121048e3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.291928] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3c4ba62e-0cd1-441f-bc9a-966fa2cf9b11 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Lock "4de32ee6-e205-435b-a75e-94c3121048e3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.292119] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3c4ba62e-0cd1-441f-bc9a-966fa2cf9b11 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Lock "4de32ee6-e205-435b-a75e-94c3121048e3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.296803] env[61868]: INFO nova.compute.manager [None req-3c4ba62e-0cd1-441f-bc9a-966fa2cf9b11 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Terminating instance [ 777.298659] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3c4ba62e-0cd1-441f-bc9a-966fa2cf9b11 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Acquiring lock "refresh_cache-4de32ee6-e205-435b-a75e-94c3121048e3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.298976] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3c4ba62e-0cd1-441f-bc9a-966fa2cf9b11 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Acquired lock "refresh_cache-4de32ee6-e205-435b-a75e-94c3121048e3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.298976] env[61868]: DEBUG nova.network.neutron [None req-3c4ba62e-0cd1-441f-bc9a-966fa2cf9b11 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 777.428331] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.399s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.429042] env[61868]: DEBUG nova.compute.manager [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 777.432984] env[61868]: DEBUG oslo_concurrency.lockutils [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.355s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.490994] env[61868]: DEBUG nova.network.neutron [-] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.816654] env[61868]: DEBUG nova.network.neutron [None req-3c4ba62e-0cd1-441f-bc9a-966fa2cf9b11 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 777.861890] env[61868]: DEBUG nova.network.neutron [None req-3c4ba62e-0cd1-441f-bc9a-966fa2cf9b11 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.940567] env[61868]: DEBUG nova.compute.utils [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 777.942912] env[61868]: DEBUG nova.compute.manager [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 777.943105] env[61868]: DEBUG nova.network.neutron [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 777.987302] env[61868]: DEBUG nova.policy [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f73bfcdddbd44790aee8d521241f1991', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e313c27c7d7342e3bc6501653c0c685d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 777.993279] env[61868]: INFO nova.compute.manager [-] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Took 1.02 seconds to deallocate network for instance. [ 777.996198] env[61868]: DEBUG nova.compute.claims [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 777.996374] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 778.201994] env[61868]: INFO nova.scheduler.client.report [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Deleted allocations for instance 2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3 [ 778.314381] env[61868]: DEBUG nova.network.neutron [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Successfully created port: 4aeee88f-b60a-47e1-bfde-6d563b574608 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 778.329863] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f33efbe-5aba-4f14-8921-d0cb779ca74b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.338077] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b30c9300-6b44-43a6-96f9-1587667aecd1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.369305] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3c4ba62e-0cd1-441f-bc9a-966fa2cf9b11 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Releasing lock "refresh_cache-4de32ee6-e205-435b-a75e-94c3121048e3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.369742] env[61868]: DEBUG nova.compute.manager [None req-3c4ba62e-0cd1-441f-bc9a-966fa2cf9b11 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 778.369941] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-3c4ba62e-0cd1-441f-bc9a-966fa2cf9b11 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 778.371369] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8e2b73c-b2af-42f6-8dad-befd7ebb4cab {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.377082] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13b6fea6-1ec0-4d1a-8411-8860f5cc56f4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.382123] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c4ba62e-0cd1-441f-bc9a-966fa2cf9b11 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 778.384164] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-835ec289-e20c-40df-a880-14a0d2311951 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.386703] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b61d6b8-33e1-46cb-8aec-91b99a657c82 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.402086] env[61868]: DEBUG nova.compute.provider_tree [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 778.407020] env[61868]: DEBUG oslo_vmware.api [None req-3c4ba62e-0cd1-441f-bc9a-966fa2cf9b11 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Waiting for the task: (returnval){ [ 778.407020] env[61868]: value = "task-1315336" [ 778.407020] env[61868]: _type = "Task" [ 778.407020] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.413786] env[61868]: DEBUG oslo_vmware.api [None req-3c4ba62e-0cd1-441f-bc9a-966fa2cf9b11 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': task-1315336, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.443964] env[61868]: DEBUG nova.compute.manager [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 778.710475] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f5616ca7-bf3a-4e15-91ee-638996f50408 tempest-ListServerFiltersTestJSON-1034861033 tempest-ListServerFiltersTestJSON-1034861033-project-member] Lock "2fabe67e-f3af-4dc7-bf07-f320f1e6e7e3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 154.650s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.906536] env[61868]: DEBUG nova.scheduler.client.report [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 778.920968] env[61868]: DEBUG oslo_vmware.api [None req-3c4ba62e-0cd1-441f-bc9a-966fa2cf9b11 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': task-1315336, 'name': PowerOffVM_Task, 'duration_secs': 0.193034} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.920968] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c4ba62e-0cd1-441f-bc9a-966fa2cf9b11 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 778.920968] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-3c4ba62e-0cd1-441f-bc9a-966fa2cf9b11 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 778.920968] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-97a409d2-2b2c-4d17-8e2c-818b683c3d24 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.945387] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-3c4ba62e-0cd1-441f-bc9a-966fa2cf9b11 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 778.945605] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-3c4ba62e-0cd1-441f-bc9a-966fa2cf9b11 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 778.945778] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c4ba62e-0cd1-441f-bc9a-966fa2cf9b11 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Deleting the datastore file [datastore1] 4de32ee6-e205-435b-a75e-94c3121048e3 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 778.946036] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-63dc78a8-d40e-4575-80e5-4f17f2f453b9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.957618] env[61868]: DEBUG oslo_vmware.api [None req-3c4ba62e-0cd1-441f-bc9a-966fa2cf9b11 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Waiting for the task: (returnval){ [ 778.957618] env[61868]: value = "task-1315338" [ 778.957618] env[61868]: _type = "Task" [ 778.957618] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.971388] env[61868]: DEBUG oslo_vmware.api [None req-3c4ba62e-0cd1-441f-bc9a-966fa2cf9b11 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': task-1315338, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.158582] env[61868]: DEBUG nova.compute.manager [req-d56eb7d0-dfbb-483d-9b33-00e5006c3fef req-68c65b45-2be1-4995-813f-63476a1da076 service nova] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Received event network-changed-4aeee88f-b60a-47e1-bfde-6d563b574608 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 779.158674] env[61868]: DEBUG nova.compute.manager [req-d56eb7d0-dfbb-483d-9b33-00e5006c3fef req-68c65b45-2be1-4995-813f-63476a1da076 service nova] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Refreshing instance network info cache due to event network-changed-4aeee88f-b60a-47e1-bfde-6d563b574608. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 779.158910] env[61868]: DEBUG oslo_concurrency.lockutils [req-d56eb7d0-dfbb-483d-9b33-00e5006c3fef req-68c65b45-2be1-4995-813f-63476a1da076 service nova] Acquiring lock "refresh_cache-6391addb-b1a5-4e15-9bd2-531eb9956b42" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.159073] env[61868]: DEBUG oslo_concurrency.lockutils [req-d56eb7d0-dfbb-483d-9b33-00e5006c3fef req-68c65b45-2be1-4995-813f-63476a1da076 service nova] Acquired lock "refresh_cache-6391addb-b1a5-4e15-9bd2-531eb9956b42" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.159242] env[61868]: DEBUG nova.network.neutron [req-d56eb7d0-dfbb-483d-9b33-00e5006c3fef req-68c65b45-2be1-4995-813f-63476a1da076 service nova] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Refreshing network info cache for port 4aeee88f-b60a-47e1-bfde-6d563b574608 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 779.203060] env[61868]: ERROR nova.compute.manager [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4aeee88f-b60a-47e1-bfde-6d563b574608, please check neutron logs for more information. [ 779.203060] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 779.203060] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 779.203060] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 779.203060] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 779.203060] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 779.203060] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 779.203060] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 779.203060] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 779.203060] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 779.203060] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 779.203060] env[61868]: ERROR nova.compute.manager raise self.value [ 779.203060] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 779.203060] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 779.203060] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 779.203060] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 779.203645] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 779.203645] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 779.203645] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4aeee88f-b60a-47e1-bfde-6d563b574608, please check neutron logs for more information. [ 779.203645] env[61868]: ERROR nova.compute.manager [ 779.203645] env[61868]: Traceback (most recent call last): [ 779.203645] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 779.203645] env[61868]: listener.cb(fileno) [ 779.203645] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 779.203645] env[61868]: result = function(*args, **kwargs) [ 779.203645] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 779.203645] env[61868]: return func(*args, **kwargs) [ 779.203645] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 779.203645] env[61868]: raise e [ 779.203645] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 779.203645] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 779.203645] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 779.203645] env[61868]: created_port_ids = self._update_ports_for_instance( [ 779.203645] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 779.203645] env[61868]: with excutils.save_and_reraise_exception(): [ 779.203645] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 779.203645] env[61868]: self.force_reraise() [ 779.203645] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 779.203645] env[61868]: raise self.value [ 779.203645] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 779.203645] env[61868]: updated_port = self._update_port( [ 779.203645] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 779.203645] env[61868]: _ensure_no_port_binding_failure(port) [ 779.203645] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 779.203645] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 779.204646] env[61868]: nova.exception.PortBindingFailed: Binding failed for port 4aeee88f-b60a-47e1-bfde-6d563b574608, please check neutron logs for more information. [ 779.204646] env[61868]: Removing descriptor: 15 [ 779.212950] env[61868]: DEBUG nova.compute.manager [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 779.414522] env[61868]: DEBUG oslo_concurrency.lockutils [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.983s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.415165] env[61868]: ERROR nova.compute.manager [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] [instance: df927c0c-b742-4451-9439-60ea63287b99] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 12eed284-8b74-4be9-ab30-4d3e30c0d76f, please check neutron logs for more information. [ 779.415165] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] Traceback (most recent call last): [ 779.415165] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 779.415165] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] self.driver.spawn(context, instance, image_meta, [ 779.415165] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 779.415165] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] self._vmops.spawn(context, instance, image_meta, injected_files, [ 779.415165] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 779.415165] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] vm_ref = self.build_virtual_machine(instance, [ 779.415165] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 779.415165] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] vif_infos = vmwarevif.get_vif_info(self._session, [ 779.415165] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 779.415571] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] for vif in network_info: [ 779.415571] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 779.415571] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] return self._sync_wrapper(fn, *args, **kwargs) [ 779.415571] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 779.415571] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] self.wait() [ 779.415571] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 779.415571] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] self[:] = self._gt.wait() [ 779.415571] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 779.415571] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] return self._exit_event.wait() [ 779.415571] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 779.415571] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] current.throw(*self._exc) [ 779.415571] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 779.415571] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] result = function(*args, **kwargs) [ 779.415969] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 779.415969] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] return func(*args, **kwargs) [ 779.415969] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 779.415969] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] raise e [ 779.415969] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 779.415969] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] nwinfo = self.network_api.allocate_for_instance( [ 779.415969] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 779.415969] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] created_port_ids = self._update_ports_for_instance( [ 779.415969] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 779.415969] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] with excutils.save_and_reraise_exception(): [ 779.415969] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 779.415969] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] self.force_reraise() [ 779.415969] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 779.416337] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] raise self.value [ 779.416337] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 779.416337] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] updated_port = self._update_port( [ 779.416337] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 779.416337] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] _ensure_no_port_binding_failure(port) [ 779.416337] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 779.416337] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] raise exception.PortBindingFailed(port_id=port['id']) [ 779.416337] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] nova.exception.PortBindingFailed: Binding failed for port 12eed284-8b74-4be9-ab30-4d3e30c0d76f, please check neutron logs for more information. [ 779.416337] env[61868]: ERROR nova.compute.manager [instance: df927c0c-b742-4451-9439-60ea63287b99] [ 779.416337] env[61868]: DEBUG nova.compute.utils [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] [instance: df927c0c-b742-4451-9439-60ea63287b99] Binding failed for port 12eed284-8b74-4be9-ab30-4d3e30c0d76f, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 779.417804] env[61868]: DEBUG nova.compute.manager [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] [instance: df927c0c-b742-4451-9439-60ea63287b99] Build of instance df927c0c-b742-4451-9439-60ea63287b99 was re-scheduled: Binding failed for port 12eed284-8b74-4be9-ab30-4d3e30c0d76f, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 779.418035] env[61868]: DEBUG nova.compute.manager [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] [instance: df927c0c-b742-4451-9439-60ea63287b99] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 779.418290] env[61868]: DEBUG oslo_concurrency.lockutils [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] Acquiring lock "refresh_cache-df927c0c-b742-4451-9439-60ea63287b99" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.418483] env[61868]: DEBUG oslo_concurrency.lockutils [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] Acquired lock "refresh_cache-df927c0c-b742-4451-9439-60ea63287b99" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.418678] env[61868]: DEBUG nova.network.neutron [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] [instance: df927c0c-b742-4451-9439-60ea63287b99] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 779.420757] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 18.835s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.420757] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.421500] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61868) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 779.421500] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.815s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.422918] env[61868]: INFO nova.compute.claims [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 779.427227] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16bb14f5-5bdb-4bae-b9fa-1a57d2e997e3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.437042] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd5a5779-d30b-4dfb-8a40-671d11fbe72b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.452059] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0308dbc1-44db-43d7-86e1-92374ba62cea {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.455638] env[61868]: DEBUG nova.compute.manager [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 779.465871] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ec20fdd-e652-4d2f-9c39-15bf14ee30a7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.475317] env[61868]: DEBUG oslo_vmware.api [None req-3c4ba62e-0cd1-441f-bc9a-966fa2cf9b11 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Task: {'id': task-1315338, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.101683} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.499512] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c4ba62e-0cd1-441f-bc9a-966fa2cf9b11 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 779.499735] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-3c4ba62e-0cd1-441f-bc9a-966fa2cf9b11 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 779.499910] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-3c4ba62e-0cd1-441f-bc9a-966fa2cf9b11 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 779.500095] env[61868]: INFO nova.compute.manager [None req-3c4ba62e-0cd1-441f-bc9a-966fa2cf9b11 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Took 1.13 seconds to destroy the instance on the hypervisor. [ 779.500350] env[61868]: DEBUG oslo.service.loopingcall [None req-3c4ba62e-0cd1-441f-bc9a-966fa2cf9b11 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 779.500975] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181524MB free_disk=134GB free_vcpus=48 pci_devices=None {{(pid=61868) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 779.501191] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.504158] env[61868]: DEBUG nova.compute.manager [-] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 779.504158] env[61868]: DEBUG nova.network.neutron [-] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 779.513020] env[61868]: DEBUG nova.virt.hardware [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 779.513020] env[61868]: DEBUG nova.virt.hardware [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 779.513020] env[61868]: DEBUG nova.virt.hardware [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 779.513327] env[61868]: DEBUG nova.virt.hardware [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 779.513327] env[61868]: DEBUG nova.virt.hardware [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 779.513327] env[61868]: DEBUG nova.virt.hardware [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 779.513327] env[61868]: DEBUG nova.virt.hardware [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 779.513327] env[61868]: DEBUG nova.virt.hardware [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 779.513511] env[61868]: DEBUG nova.virt.hardware [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 779.513511] env[61868]: DEBUG nova.virt.hardware [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 779.513511] env[61868]: DEBUG nova.virt.hardware [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 779.517016] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42745a55-ce0e-49fd-b45f-e7b8ea2b72ed {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.522281] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa4f8ca1-c299-4f3c-9b10-bc849b218513 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.527122] env[61868]: DEBUG nova.network.neutron [-] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 779.539883] env[61868]: ERROR nova.compute.manager [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4aeee88f-b60a-47e1-bfde-6d563b574608, please check neutron logs for more information. [ 779.539883] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Traceback (most recent call last): [ 779.539883] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 779.539883] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] yield resources [ 779.539883] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 779.539883] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] self.driver.spawn(context, instance, image_meta, [ 779.539883] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 779.539883] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] self._vmops.spawn(context, instance, image_meta, injected_files, [ 779.539883] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 779.539883] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] vm_ref = self.build_virtual_machine(instance, [ 779.539883] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 779.540363] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] vif_infos = vmwarevif.get_vif_info(self._session, [ 779.540363] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 779.540363] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] for vif in network_info: [ 779.540363] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 779.540363] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] return self._sync_wrapper(fn, *args, **kwargs) [ 779.540363] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 779.540363] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] self.wait() [ 779.540363] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 779.540363] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] self[:] = self._gt.wait() [ 779.540363] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 779.540363] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] return self._exit_event.wait() [ 779.540363] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 779.540363] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] current.throw(*self._exc) [ 779.540786] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 779.540786] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] result = function(*args, **kwargs) [ 779.540786] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 779.540786] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] return func(*args, **kwargs) [ 779.540786] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 779.540786] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] raise e [ 779.540786] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 779.540786] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] nwinfo = self.network_api.allocate_for_instance( [ 779.540786] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 779.540786] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] created_port_ids = self._update_ports_for_instance( [ 779.540786] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 779.540786] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] with excutils.save_and_reraise_exception(): [ 779.540786] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 779.541267] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] self.force_reraise() [ 779.541267] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 779.541267] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] raise self.value [ 779.541267] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 779.541267] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] updated_port = self._update_port( [ 779.541267] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 779.541267] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] _ensure_no_port_binding_failure(port) [ 779.541267] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 779.541267] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] raise exception.PortBindingFailed(port_id=port['id']) [ 779.541267] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] nova.exception.PortBindingFailed: Binding failed for port 4aeee88f-b60a-47e1-bfde-6d563b574608, please check neutron logs for more information. [ 779.541267] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] [ 779.541267] env[61868]: INFO nova.compute.manager [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Terminating instance [ 779.542109] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Acquiring lock "refresh_cache-6391addb-b1a5-4e15-9bd2-531eb9956b42" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.678593] env[61868]: DEBUG nova.network.neutron [req-d56eb7d0-dfbb-483d-9b33-00e5006c3fef req-68c65b45-2be1-4995-813f-63476a1da076 service nova] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 779.734288] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.769659] env[61868]: DEBUG nova.network.neutron [req-d56eb7d0-dfbb-483d-9b33-00e5006c3fef req-68c65b45-2be1-4995-813f-63476a1da076 service nova] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.936176] env[61868]: DEBUG nova.network.neutron [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] [instance: df927c0c-b742-4451-9439-60ea63287b99] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 780.013145] env[61868]: DEBUG nova.network.neutron [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] [instance: df927c0c-b742-4451-9439-60ea63287b99] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.029937] env[61868]: DEBUG nova.network.neutron [-] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.272585] env[61868]: DEBUG oslo_concurrency.lockutils [req-d56eb7d0-dfbb-483d-9b33-00e5006c3fef req-68c65b45-2be1-4995-813f-63476a1da076 service nova] Releasing lock "refresh_cache-6391addb-b1a5-4e15-9bd2-531eb9956b42" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.272996] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Acquired lock "refresh_cache-6391addb-b1a5-4e15-9bd2-531eb9956b42" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.273228] env[61868]: DEBUG nova.network.neutron [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 780.516263] env[61868]: DEBUG oslo_concurrency.lockutils [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] Releasing lock "refresh_cache-df927c0c-b742-4451-9439-60ea63287b99" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.516504] env[61868]: DEBUG nova.compute.manager [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 780.516705] env[61868]: DEBUG nova.compute.manager [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] [instance: df927c0c-b742-4451-9439-60ea63287b99] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 780.516938] env[61868]: DEBUG nova.network.neutron [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] [instance: df927c0c-b742-4451-9439-60ea63287b99] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 780.533282] env[61868]: DEBUG nova.network.neutron [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] [instance: df927c0c-b742-4451-9439-60ea63287b99] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 780.534447] env[61868]: INFO nova.compute.manager [-] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Took 1.03 seconds to deallocate network for instance. [ 780.743108] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-682d09c4-b9d1-4772-80ce-b1fa79f0c063 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.751091] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb2429de-e0d0-44e6-8cd8-965d368e2018 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.782589] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f06d013-8cf3-4697-a8e0-0552008ccc21 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.790387] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f727b43-eb7d-4aaa-a6e7-e099617cdfa9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.804259] env[61868]: DEBUG nova.compute.provider_tree [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 780.806258] env[61868]: DEBUG nova.network.neutron [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 780.879847] env[61868]: DEBUG nova.network.neutron [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.035489] env[61868]: DEBUG nova.network.neutron [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] [instance: df927c0c-b742-4451-9439-60ea63287b99] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.039976] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3c4ba62e-0cd1-441f-bc9a-966fa2cf9b11 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.180995] env[61868]: DEBUG nova.compute.manager [req-9e954ca7-cf58-4f3a-9df0-625266e4a4dd req-b813d7a8-76a3-46ce-8dc3-0909786d6fb1 service nova] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Received event network-vif-deleted-4aeee88f-b60a-47e1-bfde-6d563b574608 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 781.309902] env[61868]: DEBUG nova.scheduler.client.report [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 781.382812] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Releasing lock "refresh_cache-6391addb-b1a5-4e15-9bd2-531eb9956b42" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.383288] env[61868]: DEBUG nova.compute.manager [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 781.383483] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 781.383764] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9a8d86a8-adb8-40c2-bd33-1f5719b3a6a2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.394511] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e81e8a4c-2843-466e-824b-85c92d6ac2c0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.412809] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6391addb-b1a5-4e15-9bd2-531eb9956b42 could not be found. [ 781.412971] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 781.413891] env[61868]: INFO nova.compute.manager [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Took 0.03 seconds to destroy the instance on the hypervisor. [ 781.413891] env[61868]: DEBUG oslo.service.loopingcall [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 781.413891] env[61868]: DEBUG nova.compute.manager [-] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 781.413891] env[61868]: DEBUG nova.network.neutron [-] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 781.428936] env[61868]: DEBUG nova.network.neutron [-] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 781.538227] env[61868]: INFO nova.compute.manager [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] [instance: df927c0c-b742-4451-9439-60ea63287b99] Took 1.02 seconds to deallocate network for instance. [ 781.816515] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.395s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.817068] env[61868]: DEBUG nova.compute.manager [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 781.820218] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.172s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.931365] env[61868]: DEBUG nova.network.neutron [-] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.324368] env[61868]: DEBUG nova.compute.utils [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 782.325750] env[61868]: DEBUG nova.compute.manager [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 782.325923] env[61868]: DEBUG nova.network.neutron [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 782.368135] env[61868]: DEBUG nova.policy [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'daad105e7edf4fb0ae0b2e685bfd92e9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b520c3ac58074e8d9b0bfafb817244a8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 782.433902] env[61868]: INFO nova.compute.manager [-] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Took 1.02 seconds to deallocate network for instance. [ 782.436549] env[61868]: DEBUG nova.compute.claims [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 782.436739] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.570582] env[61868]: INFO nova.scheduler.client.report [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] Deleted allocations for instance df927c0c-b742-4451-9439-60ea63287b99 [ 782.678022] env[61868]: DEBUG nova.network.neutron [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Successfully created port: c99d18ba-2fcd-4d59-a606-7c99be343788 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 782.705511] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aefbe72-cb15-4861-bb53-0aed0817b9cd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.712998] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-269da84e-223a-409f-9433-9f45b5bfc2ad {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.744731] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fd0b7bd-47c2-4e58-9d25-06ad8b05293f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.754024] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06e43475-e5c7-43fb-8879-413f581ef811 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.764994] env[61868]: DEBUG nova.compute.provider_tree [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 782.834409] env[61868]: DEBUG nova.compute.manager [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 783.081200] env[61868]: DEBUG oslo_concurrency.lockutils [None req-35b49efc-51de-47b3-9768-e7a1c9548572 tempest-ServersNegativeTestMultiTenantJSON-19489647 tempest-ServersNegativeTestMultiTenantJSON-19489647-project-member] Lock "df927c0c-b742-4451-9439-60ea63287b99" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 158.406s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.271924] env[61868]: DEBUG nova.scheduler.client.report [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 783.404863] env[61868]: DEBUG nova.compute.manager [req-37d7d7b1-3eea-4435-a3b6-6a70de24992c req-81207213-ea2f-4618-8ee0-24eda9cf6c1e service nova] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Received event network-changed-c99d18ba-2fcd-4d59-a606-7c99be343788 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 783.404863] env[61868]: DEBUG nova.compute.manager [req-37d7d7b1-3eea-4435-a3b6-6a70de24992c req-81207213-ea2f-4618-8ee0-24eda9cf6c1e service nova] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Refreshing instance network info cache due to event network-changed-c99d18ba-2fcd-4d59-a606-7c99be343788. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 783.404863] env[61868]: DEBUG oslo_concurrency.lockutils [req-37d7d7b1-3eea-4435-a3b6-6a70de24992c req-81207213-ea2f-4618-8ee0-24eda9cf6c1e service nova] Acquiring lock "refresh_cache-c0be3a6b-4d9d-4841-b05c-6aca4662b004" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.405521] env[61868]: DEBUG oslo_concurrency.lockutils [req-37d7d7b1-3eea-4435-a3b6-6a70de24992c req-81207213-ea2f-4618-8ee0-24eda9cf6c1e service nova] Acquired lock "refresh_cache-c0be3a6b-4d9d-4841-b05c-6aca4662b004" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.405521] env[61868]: DEBUG nova.network.neutron [req-37d7d7b1-3eea-4435-a3b6-6a70de24992c req-81207213-ea2f-4618-8ee0-24eda9cf6c1e service nova] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Refreshing network info cache for port c99d18ba-2fcd-4d59-a606-7c99be343788 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 783.572793] env[61868]: ERROR nova.compute.manager [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c99d18ba-2fcd-4d59-a606-7c99be343788, please check neutron logs for more information. [ 783.572793] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 783.572793] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 783.572793] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 783.572793] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 783.572793] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 783.572793] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 783.572793] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 783.572793] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 783.572793] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 783.572793] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 783.572793] env[61868]: ERROR nova.compute.manager raise self.value [ 783.572793] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 783.572793] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 783.572793] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 783.572793] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 783.573373] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 783.573373] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 783.573373] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c99d18ba-2fcd-4d59-a606-7c99be343788, please check neutron logs for more information. [ 783.573373] env[61868]: ERROR nova.compute.manager [ 783.573373] env[61868]: Traceback (most recent call last): [ 783.574169] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 783.574169] env[61868]: listener.cb(fileno) [ 783.574169] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 783.574169] env[61868]: result = function(*args, **kwargs) [ 783.574169] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 783.574169] env[61868]: return func(*args, **kwargs) [ 783.574169] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 783.574169] env[61868]: raise e [ 783.574454] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 783.574454] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 783.574454] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 783.574454] env[61868]: created_port_ids = self._update_ports_for_instance( [ 783.574454] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 783.574454] env[61868]: with excutils.save_and_reraise_exception(): [ 783.574454] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 783.574454] env[61868]: self.force_reraise() [ 783.574454] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 783.574454] env[61868]: raise self.value [ 783.574454] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 783.574454] env[61868]: updated_port = self._update_port( [ 783.574454] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 783.574454] env[61868]: _ensure_no_port_binding_failure(port) [ 783.574454] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 783.574454] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 783.574454] env[61868]: nova.exception.PortBindingFailed: Binding failed for port c99d18ba-2fcd-4d59-a606-7c99be343788, please check neutron logs for more information. [ 783.574454] env[61868]: Removing descriptor: 15 [ 783.584086] env[61868]: DEBUG nova.compute.manager [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 783.778052] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.958s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.778691] env[61868]: ERROR nova.compute.manager [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1ef56e50-9ea3-48fb-b301-847f27e579b0, please check neutron logs for more information. [ 783.778691] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Traceback (most recent call last): [ 783.778691] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 783.778691] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] self.driver.spawn(context, instance, image_meta, [ 783.778691] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 783.778691] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 783.778691] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 783.778691] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] vm_ref = self.build_virtual_machine(instance, [ 783.778691] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 783.778691] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] vif_infos = vmwarevif.get_vif_info(self._session, [ 783.778691] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 783.779070] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] for vif in network_info: [ 783.779070] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 783.779070] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] return self._sync_wrapper(fn, *args, **kwargs) [ 783.779070] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 783.779070] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] self.wait() [ 783.779070] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 783.779070] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] self[:] = self._gt.wait() [ 783.779070] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 783.779070] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] return self._exit_event.wait() [ 783.779070] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 783.779070] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] current.throw(*self._exc) [ 783.779070] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 783.779070] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] result = function(*args, **kwargs) [ 783.779506] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 783.779506] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] return func(*args, **kwargs) [ 783.779506] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 783.779506] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] raise e [ 783.779506] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 783.779506] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] nwinfo = self.network_api.allocate_for_instance( [ 783.779506] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 783.779506] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] created_port_ids = self._update_ports_for_instance( [ 783.779506] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 783.779506] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] with excutils.save_and_reraise_exception(): [ 783.779506] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 783.779506] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] self.force_reraise() [ 783.779506] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 783.779863] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] raise self.value [ 783.779863] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 783.779863] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] updated_port = self._update_port( [ 783.779863] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 783.779863] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] _ensure_no_port_binding_failure(port) [ 783.779863] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 783.779863] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] raise exception.PortBindingFailed(port_id=port['id']) [ 783.779863] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] nova.exception.PortBindingFailed: Binding failed for port 1ef56e50-9ea3-48fb-b301-847f27e579b0, please check neutron logs for more information. [ 783.779863] env[61868]: ERROR nova.compute.manager [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] [ 783.779863] env[61868]: DEBUG nova.compute.utils [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Binding failed for port 1ef56e50-9ea3-48fb-b301-847f27e579b0, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 783.780993] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.605s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.782544] env[61868]: INFO nova.compute.claims [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 783.785049] env[61868]: DEBUG nova.compute.manager [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Build of instance a0fda7cf-21b6-4a22-bf56-09e848b4dc9e was re-scheduled: Binding failed for port 1ef56e50-9ea3-48fb-b301-847f27e579b0, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 783.785523] env[61868]: DEBUG nova.compute.manager [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 783.785684] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] Acquiring lock "refresh_cache-a0fda7cf-21b6-4a22-bf56-09e848b4dc9e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.785830] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] Acquired lock "refresh_cache-a0fda7cf-21b6-4a22-bf56-09e848b4dc9e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.785986] env[61868]: DEBUG nova.network.neutron [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 783.846484] env[61868]: DEBUG nova.compute.manager [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 783.871298] env[61868]: DEBUG nova.virt.hardware [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 783.873017] env[61868]: DEBUG nova.virt.hardware [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 783.873017] env[61868]: DEBUG nova.virt.hardware [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 783.873017] env[61868]: DEBUG nova.virt.hardware [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 783.873017] env[61868]: DEBUG nova.virt.hardware [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 783.873017] env[61868]: DEBUG nova.virt.hardware [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 783.873017] env[61868]: DEBUG nova.virt.hardware [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 783.873259] env[61868]: DEBUG nova.virt.hardware [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 783.873259] env[61868]: DEBUG nova.virt.hardware [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 783.873622] env[61868]: DEBUG nova.virt.hardware [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 783.877016] env[61868]: DEBUG nova.virt.hardware [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 783.877016] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fbbc5af-a2d1-4b01-addc-04883788e03a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.883764] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b6f01a7-25e7-460b-a7f2-d895a72b1f08 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.898813] env[61868]: ERROR nova.compute.manager [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c99d18ba-2fcd-4d59-a606-7c99be343788, please check neutron logs for more information. [ 783.898813] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Traceback (most recent call last): [ 783.898813] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 783.898813] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] yield resources [ 783.898813] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 783.898813] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] self.driver.spawn(context, instance, image_meta, [ 783.898813] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 783.898813] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] self._vmops.spawn(context, instance, image_meta, injected_files, [ 783.898813] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 783.898813] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] vm_ref = self.build_virtual_machine(instance, [ 783.898813] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 783.899280] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] vif_infos = vmwarevif.get_vif_info(self._session, [ 783.899280] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 783.899280] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] for vif in network_info: [ 783.899280] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 783.899280] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] return self._sync_wrapper(fn, *args, **kwargs) [ 783.899280] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 783.899280] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] self.wait() [ 783.899280] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 783.899280] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] self[:] = self._gt.wait() [ 783.899280] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 783.899280] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] return self._exit_event.wait() [ 783.899280] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 783.899280] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] current.throw(*self._exc) [ 783.899656] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 783.899656] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] result = function(*args, **kwargs) [ 783.899656] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 783.899656] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] return func(*args, **kwargs) [ 783.899656] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 783.899656] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] raise e [ 783.899656] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 783.899656] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] nwinfo = self.network_api.allocate_for_instance( [ 783.899656] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 783.899656] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] created_port_ids = self._update_ports_for_instance( [ 783.899656] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 783.899656] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] with excutils.save_and_reraise_exception(): [ 783.899656] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 783.900041] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] self.force_reraise() [ 783.900041] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 783.900041] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] raise self.value [ 783.900041] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 783.900041] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] updated_port = self._update_port( [ 783.900041] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 783.900041] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] _ensure_no_port_binding_failure(port) [ 783.900041] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 783.900041] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] raise exception.PortBindingFailed(port_id=port['id']) [ 783.900041] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] nova.exception.PortBindingFailed: Binding failed for port c99d18ba-2fcd-4d59-a606-7c99be343788, please check neutron logs for more information. [ 783.900041] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] [ 783.901057] env[61868]: INFO nova.compute.manager [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Terminating instance [ 783.907179] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "refresh_cache-c0be3a6b-4d9d-4841-b05c-6aca4662b004" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.931462] env[61868]: DEBUG nova.network.neutron [req-37d7d7b1-3eea-4435-a3b6-6a70de24992c req-81207213-ea2f-4618-8ee0-24eda9cf6c1e service nova] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 784.058458] env[61868]: DEBUG nova.network.neutron [req-37d7d7b1-3eea-4435-a3b6-6a70de24992c req-81207213-ea2f-4618-8ee0-24eda9cf6c1e service nova] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.105677] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.308147] env[61868]: DEBUG nova.network.neutron [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 784.420136] env[61868]: DEBUG nova.network.neutron [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.564609] env[61868]: DEBUG oslo_concurrency.lockutils [req-37d7d7b1-3eea-4435-a3b6-6a70de24992c req-81207213-ea2f-4618-8ee0-24eda9cf6c1e service nova] Releasing lock "refresh_cache-c0be3a6b-4d9d-4841-b05c-6aca4662b004" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.565374] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquired lock "refresh_cache-c0be3a6b-4d9d-4841-b05c-6aca4662b004" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.565774] env[61868]: DEBUG nova.network.neutron [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 784.922783] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] Releasing lock "refresh_cache-a0fda7cf-21b6-4a22-bf56-09e848b4dc9e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.923026] env[61868]: DEBUG nova.compute.manager [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 784.923255] env[61868]: DEBUG nova.compute.manager [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 784.923410] env[61868]: DEBUG nova.network.neutron [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 784.944436] env[61868]: DEBUG nova.network.neutron [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 785.093360] env[61868]: DEBUG nova.network.neutron [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 785.101768] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63074a72-3f2c-4b8b-a18c-f3dcaca332b7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.109091] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9978d349-dfb3-4ba9-b97b-216b799214b5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.138428] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3018492-b122-40b6-b582-606fd9cead2a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.145176] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca24746d-8c82-45a7-a327-0f8ea0c3681b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.159036] env[61868]: DEBUG nova.compute.provider_tree [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 785.220539] env[61868]: DEBUG nova.network.neutron [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.447337] env[61868]: DEBUG nova.network.neutron [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.477543] env[61868]: DEBUG nova.compute.manager [req-389b2156-9c46-43cb-be41-52a5f1240005 req-ec955042-36d9-4d88-89bf-2b87bba254ba service nova] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Received event network-vif-deleted-c99d18ba-2fcd-4d59-a606-7c99be343788 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 785.665111] env[61868]: DEBUG nova.scheduler.client.report [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 785.725177] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Releasing lock "refresh_cache-c0be3a6b-4d9d-4841-b05c-6aca4662b004" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.725615] env[61868]: DEBUG nova.compute.manager [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 785.725842] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 785.726143] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6fabb1d1-9857-4226-8191-0576aaecc836 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.734898] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99d26a30-44a9-4824-a489-49d8764a54d0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.756887] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c0be3a6b-4d9d-4841-b05c-6aca4662b004 could not be found. [ 785.757102] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 785.757278] env[61868]: INFO nova.compute.manager [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Took 0.03 seconds to destroy the instance on the hypervisor. [ 785.757592] env[61868]: DEBUG oslo.service.loopingcall [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 785.757713] env[61868]: DEBUG nova.compute.manager [-] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 785.757843] env[61868]: DEBUG nova.network.neutron [-] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 785.775786] env[61868]: DEBUG nova.network.neutron [-] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 785.951623] env[61868]: INFO nova.compute.manager [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] [instance: a0fda7cf-21b6-4a22-bf56-09e848b4dc9e] Took 1.03 seconds to deallocate network for instance. [ 786.167567] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.386s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.170339] env[61868]: DEBUG oslo_concurrency.lockutils [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.663s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.283529] env[61868]: DEBUG nova.network.neutron [-] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.671512] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Acquiring lock "f34b0980-2855-4778-aad4-9266b0f9d862" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.671889] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Lock "f34b0980-2855-4778-aad4-9266b0f9d862" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.786148] env[61868]: INFO nova.compute.manager [-] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Took 1.03 seconds to deallocate network for instance. [ 786.788932] env[61868]: DEBUG nova.compute.claims [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 786.789130] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.956308] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1df4fe33-79af-4a23-8f9f-50779f9b2e22 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.963968] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f6bba86-a87d-4074-9983-966bfc8100b1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.995089] env[61868]: INFO nova.scheduler.client.report [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] Deleted allocations for instance a0fda7cf-21b6-4a22-bf56-09e848b4dc9e [ 787.000815] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bee07b3f-d620-4c06-bb30-df0177f04c5d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.008779] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77215388-5cdd-4927-af90-654d180cf395 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.021704] env[61868]: DEBUG nova.compute.provider_tree [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 787.174141] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Lock "f34b0980-2855-4778-aad4-9266b0f9d862" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.502s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.174654] env[61868]: DEBUG nova.compute.manager [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 787.505522] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6139154c-a2a0-4fb0-9351-953a668b4885 tempest-InstanceActionsNegativeTestJSON-1689069949 tempest-InstanceActionsNegativeTestJSON-1689069949-project-member] Lock "a0fda7cf-21b6-4a22-bf56-09e848b4dc9e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 156.384s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.527987] env[61868]: DEBUG nova.scheduler.client.report [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 787.679507] env[61868]: DEBUG nova.compute.utils [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 787.680531] env[61868]: DEBUG nova.compute.manager [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 787.680704] env[61868]: DEBUG nova.network.neutron [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 787.737173] env[61868]: DEBUG nova.policy [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cbe5c7c308204cc9af6e9aa91896d5cc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '170de1ec8be64e1abcef4cf93533bedc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 787.988124] env[61868]: DEBUG nova.network.neutron [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Successfully created port: 6b4cdaf4-10aa-4932-b8ad-9c3d42d2d9cf {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 788.008191] env[61868]: DEBUG nova.compute.manager [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 788.032464] env[61868]: DEBUG oslo_concurrency.lockutils [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.862s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.033122] env[61868]: ERROR nova.compute.manager [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a9b0a8d5-d2f4-4b81-952f-f194085c3b38, please check neutron logs for more information. [ 788.033122] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Traceback (most recent call last): [ 788.033122] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 788.033122] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] self.driver.spawn(context, instance, image_meta, [ 788.033122] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 788.033122] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] self._vmops.spawn(context, instance, image_meta, injected_files, [ 788.033122] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 788.033122] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] vm_ref = self.build_virtual_machine(instance, [ 788.033122] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 788.033122] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] vif_infos = vmwarevif.get_vif_info(self._session, [ 788.033122] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 788.033502] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] for vif in network_info: [ 788.033502] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 788.033502] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] return self._sync_wrapper(fn, *args, **kwargs) [ 788.033502] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 788.033502] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] self.wait() [ 788.033502] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 788.033502] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] self[:] = self._gt.wait() [ 788.033502] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 788.033502] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] return self._exit_event.wait() [ 788.033502] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 788.033502] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] current.throw(*self._exc) [ 788.033502] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 788.033502] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] result = function(*args, **kwargs) [ 788.033945] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 788.033945] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] return func(*args, **kwargs) [ 788.033945] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 788.033945] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] raise e [ 788.033945] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 788.033945] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] nwinfo = self.network_api.allocate_for_instance( [ 788.033945] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 788.033945] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] created_port_ids = self._update_ports_for_instance( [ 788.033945] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 788.033945] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] with excutils.save_and_reraise_exception(): [ 788.033945] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 788.033945] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] self.force_reraise() [ 788.033945] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 788.034355] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] raise self.value [ 788.034355] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 788.034355] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] updated_port = self._update_port( [ 788.034355] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 788.034355] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] _ensure_no_port_binding_failure(port) [ 788.034355] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 788.034355] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] raise exception.PortBindingFailed(port_id=port['id']) [ 788.034355] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] nova.exception.PortBindingFailed: Binding failed for port a9b0a8d5-d2f4-4b81-952f-f194085c3b38, please check neutron logs for more information. [ 788.034355] env[61868]: ERROR nova.compute.manager [instance: ebb09f50-50b8-49bb-b366-5b8906605699] [ 788.034355] env[61868]: DEBUG nova.compute.utils [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Binding failed for port a9b0a8d5-d2f4-4b81-952f-f194085c3b38, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 788.034945] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.303s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.038559] env[61868]: DEBUG nova.compute.manager [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Build of instance ebb09f50-50b8-49bb-b366-5b8906605699 was re-scheduled: Binding failed for port a9b0a8d5-d2f4-4b81-952f-f194085c3b38, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 788.042500] env[61868]: DEBUG nova.compute.manager [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 788.042738] env[61868]: DEBUG oslo_concurrency.lockutils [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "refresh_cache-ebb09f50-50b8-49bb-b366-5b8906605699" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.042880] env[61868]: DEBUG oslo_concurrency.lockutils [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquired lock "refresh_cache-ebb09f50-50b8-49bb-b366-5b8906605699" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.043048] env[61868]: DEBUG nova.network.neutron [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 788.183947] env[61868]: DEBUG nova.compute.manager [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 788.534462] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.563778] env[61868]: DEBUG nova.network.neutron [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 788.675722] env[61868]: DEBUG nova.network.neutron [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.871315] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8be0dd4-b8ee-488d-89b8-0466eaf1dbf1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.879383] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bda7a40e-778f-4731-9585-86de4c770351 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.917449] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa47eedc-ad2b-49c9-bcfd-b150849f04f4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.920884] env[61868]: DEBUG nova.compute.manager [req-7686bf17-625e-46de-82dd-aec8f8f28213 req-3aa1d079-9c34-4f74-b03d-0d93a19ea7d0 service nova] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Received event network-changed-6b4cdaf4-10aa-4932-b8ad-9c3d42d2d9cf {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 788.921089] env[61868]: DEBUG nova.compute.manager [req-7686bf17-625e-46de-82dd-aec8f8f28213 req-3aa1d079-9c34-4f74-b03d-0d93a19ea7d0 service nova] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Refreshing instance network info cache due to event network-changed-6b4cdaf4-10aa-4932-b8ad-9c3d42d2d9cf. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 788.921306] env[61868]: DEBUG oslo_concurrency.lockutils [req-7686bf17-625e-46de-82dd-aec8f8f28213 req-3aa1d079-9c34-4f74-b03d-0d93a19ea7d0 service nova] Acquiring lock "refresh_cache-cd1ccbd7-5d2e-4dd9-a62e-24706759a67a" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.921444] env[61868]: DEBUG oslo_concurrency.lockutils [req-7686bf17-625e-46de-82dd-aec8f8f28213 req-3aa1d079-9c34-4f74-b03d-0d93a19ea7d0 service nova] Acquired lock "refresh_cache-cd1ccbd7-5d2e-4dd9-a62e-24706759a67a" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.921606] env[61868]: DEBUG nova.network.neutron [req-7686bf17-625e-46de-82dd-aec8f8f28213 req-3aa1d079-9c34-4f74-b03d-0d93a19ea7d0 service nova] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Refreshing network info cache for port 6b4cdaf4-10aa-4932-b8ad-9c3d42d2d9cf {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 788.928233] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4641656a-8a5e-47c6-8f35-7e9ea966d13b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.942871] env[61868]: DEBUG nova.compute.provider_tree [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 789.063271] env[61868]: ERROR nova.compute.manager [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6b4cdaf4-10aa-4932-b8ad-9c3d42d2d9cf, please check neutron logs for more information. [ 789.063271] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 789.063271] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 789.063271] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 789.063271] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 789.063271] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 789.063271] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 789.063271] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 789.063271] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 789.063271] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 789.063271] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 789.063271] env[61868]: ERROR nova.compute.manager raise self.value [ 789.063271] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 789.063271] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 789.063271] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 789.063271] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 789.064047] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 789.064047] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 789.064047] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6b4cdaf4-10aa-4932-b8ad-9c3d42d2d9cf, please check neutron logs for more information. [ 789.064047] env[61868]: ERROR nova.compute.manager [ 789.064047] env[61868]: Traceback (most recent call last): [ 789.064047] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 789.064047] env[61868]: listener.cb(fileno) [ 789.064047] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 789.064047] env[61868]: result = function(*args, **kwargs) [ 789.064047] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 789.064047] env[61868]: return func(*args, **kwargs) [ 789.064047] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 789.064047] env[61868]: raise e [ 789.064047] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 789.064047] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 789.064047] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 789.064047] env[61868]: created_port_ids = self._update_ports_for_instance( [ 789.064047] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 789.064047] env[61868]: with excutils.save_and_reraise_exception(): [ 789.064047] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 789.064047] env[61868]: self.force_reraise() [ 789.064047] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 789.064047] env[61868]: raise self.value [ 789.064047] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 789.064047] env[61868]: updated_port = self._update_port( [ 789.064047] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 789.064047] env[61868]: _ensure_no_port_binding_failure(port) [ 789.064047] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 789.064047] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 789.066095] env[61868]: nova.exception.PortBindingFailed: Binding failed for port 6b4cdaf4-10aa-4932-b8ad-9c3d42d2d9cf, please check neutron logs for more information. [ 789.066095] env[61868]: Removing descriptor: 15 [ 789.178283] env[61868]: DEBUG oslo_concurrency.lockutils [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Releasing lock "refresh_cache-ebb09f50-50b8-49bb-b366-5b8906605699" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.178696] env[61868]: DEBUG nova.compute.manager [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 789.178696] env[61868]: DEBUG nova.compute.manager [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 789.178876] env[61868]: DEBUG nova.network.neutron [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 789.194507] env[61868]: DEBUG nova.network.neutron [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 789.196569] env[61868]: DEBUG nova.compute.manager [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 789.223725] env[61868]: DEBUG nova.virt.hardware [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 789.223997] env[61868]: DEBUG nova.virt.hardware [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 789.224194] env[61868]: DEBUG nova.virt.hardware [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 789.224441] env[61868]: DEBUG nova.virt.hardware [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 789.224515] env[61868]: DEBUG nova.virt.hardware [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 789.224656] env[61868]: DEBUG nova.virt.hardware [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 789.224858] env[61868]: DEBUG nova.virt.hardware [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 789.225868] env[61868]: DEBUG nova.virt.hardware [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 789.226136] env[61868]: DEBUG nova.virt.hardware [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 789.226313] env[61868]: DEBUG nova.virt.hardware [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 789.226583] env[61868]: DEBUG nova.virt.hardware [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 789.227521] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-339c79a9-16c2-4d56-bd40-e78013edae04 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.236912] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-329d94e5-96dc-468f-9fff-50151740e955 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.252632] env[61868]: ERROR nova.compute.manager [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6b4cdaf4-10aa-4932-b8ad-9c3d42d2d9cf, please check neutron logs for more information. [ 789.252632] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Traceback (most recent call last): [ 789.252632] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 789.252632] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] yield resources [ 789.252632] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 789.252632] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] self.driver.spawn(context, instance, image_meta, [ 789.252632] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 789.252632] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 789.252632] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 789.252632] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] vm_ref = self.build_virtual_machine(instance, [ 789.252632] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 789.253158] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] vif_infos = vmwarevif.get_vif_info(self._session, [ 789.253158] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 789.253158] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] for vif in network_info: [ 789.253158] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 789.253158] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] return self._sync_wrapper(fn, *args, **kwargs) [ 789.253158] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 789.253158] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] self.wait() [ 789.253158] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 789.253158] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] self[:] = self._gt.wait() [ 789.253158] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 789.253158] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] return self._exit_event.wait() [ 789.253158] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 789.253158] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] current.throw(*self._exc) [ 789.253603] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 789.253603] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] result = function(*args, **kwargs) [ 789.253603] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 789.253603] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] return func(*args, **kwargs) [ 789.253603] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 789.253603] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] raise e [ 789.253603] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 789.253603] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] nwinfo = self.network_api.allocate_for_instance( [ 789.253603] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 789.253603] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] created_port_ids = self._update_ports_for_instance( [ 789.253603] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 789.253603] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] with excutils.save_and_reraise_exception(): [ 789.253603] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 789.254073] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] self.force_reraise() [ 789.254073] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 789.254073] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] raise self.value [ 789.254073] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 789.254073] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] updated_port = self._update_port( [ 789.254073] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 789.254073] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] _ensure_no_port_binding_failure(port) [ 789.254073] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 789.254073] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] raise exception.PortBindingFailed(port_id=port['id']) [ 789.254073] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] nova.exception.PortBindingFailed: Binding failed for port 6b4cdaf4-10aa-4932-b8ad-9c3d42d2d9cf, please check neutron logs for more information. [ 789.254073] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] [ 789.254073] env[61868]: INFO nova.compute.manager [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Terminating instance [ 789.255163] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Acquiring lock "refresh_cache-cd1ccbd7-5d2e-4dd9-a62e-24706759a67a" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.444355] env[61868]: DEBUG nova.network.neutron [req-7686bf17-625e-46de-82dd-aec8f8f28213 req-3aa1d079-9c34-4f74-b03d-0d93a19ea7d0 service nova] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 789.446662] env[61868]: DEBUG nova.scheduler.client.report [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 789.534532] env[61868]: DEBUG nova.network.neutron [req-7686bf17-625e-46de-82dd-aec8f8f28213 req-3aa1d079-9c34-4f74-b03d-0d93a19ea7d0 service nova] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.700549] env[61868]: DEBUG nova.network.neutron [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.952269] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.917s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.952919] env[61868]: ERROR nova.compute.manager [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7eec1a36-0cee-40ed-827c-749d6c1dab06, please check neutron logs for more information. [ 789.952919] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Traceback (most recent call last): [ 789.952919] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 789.952919] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] self.driver.spawn(context, instance, image_meta, [ 789.952919] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 789.952919] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 789.952919] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 789.952919] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] vm_ref = self.build_virtual_machine(instance, [ 789.952919] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 789.952919] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] vif_infos = vmwarevif.get_vif_info(self._session, [ 789.952919] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 789.953233] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] for vif in network_info: [ 789.953233] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 789.953233] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] return self._sync_wrapper(fn, *args, **kwargs) [ 789.953233] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 789.953233] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] self.wait() [ 789.953233] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 789.953233] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] self[:] = self._gt.wait() [ 789.953233] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 789.953233] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] return self._exit_event.wait() [ 789.953233] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 789.953233] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] current.throw(*self._exc) [ 789.953233] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 789.953233] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] result = function(*args, **kwargs) [ 789.953593] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 789.953593] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] return func(*args, **kwargs) [ 789.953593] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 789.953593] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] raise e [ 789.953593] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 789.953593] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] nwinfo = self.network_api.allocate_for_instance( [ 789.953593] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 789.953593] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] created_port_ids = self._update_ports_for_instance( [ 789.953593] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 789.953593] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] with excutils.save_and_reraise_exception(): [ 789.953593] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 789.953593] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] self.force_reraise() [ 789.953593] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 789.953960] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] raise self.value [ 789.953960] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 789.953960] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] updated_port = self._update_port( [ 789.953960] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 789.953960] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] _ensure_no_port_binding_failure(port) [ 789.953960] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 789.953960] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] raise exception.PortBindingFailed(port_id=port['id']) [ 789.953960] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] nova.exception.PortBindingFailed: Binding failed for port 7eec1a36-0cee-40ed-827c-749d6c1dab06, please check neutron logs for more information. [ 789.953960] env[61868]: ERROR nova.compute.manager [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] [ 789.953960] env[61868]: DEBUG nova.compute.utils [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Binding failed for port 7eec1a36-0cee-40ed-827c-749d6c1dab06, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 789.954809] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.642s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.956253] env[61868]: INFO nova.compute.claims [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 789.958877] env[61868]: DEBUG nova.compute.manager [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Build of instance 29530178-d69c-4aed-9061-d3d1cfa954d0 was re-scheduled: Binding failed for port 7eec1a36-0cee-40ed-827c-749d6c1dab06, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 789.959186] env[61868]: DEBUG nova.compute.manager [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 789.959750] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Acquiring lock "refresh_cache-29530178-d69c-4aed-9061-d3d1cfa954d0" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.959750] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Acquired lock "refresh_cache-29530178-d69c-4aed-9061-d3d1cfa954d0" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.959750] env[61868]: DEBUG nova.network.neutron [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 790.038117] env[61868]: DEBUG oslo_concurrency.lockutils [req-7686bf17-625e-46de-82dd-aec8f8f28213 req-3aa1d079-9c34-4f74-b03d-0d93a19ea7d0 service nova] Releasing lock "refresh_cache-cd1ccbd7-5d2e-4dd9-a62e-24706759a67a" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.038704] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Acquired lock "refresh_cache-cd1ccbd7-5d2e-4dd9-a62e-24706759a67a" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.038704] env[61868]: DEBUG nova.network.neutron [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 790.203189] env[61868]: INFO nova.compute.manager [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: ebb09f50-50b8-49bb-b366-5b8906605699] Took 1.02 seconds to deallocate network for instance. [ 790.481718] env[61868]: DEBUG nova.network.neutron [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 790.558996] env[61868]: DEBUG nova.network.neutron [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 790.591339] env[61868]: DEBUG nova.network.neutron [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.641391] env[61868]: DEBUG nova.network.neutron [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.786104] env[61868]: DEBUG oslo_concurrency.lockutils [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Acquiring lock "cf063dbc-7b5a-4836-91bf-a0aa33cca6bc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.786344] env[61868]: DEBUG oslo_concurrency.lockutils [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Lock "cf063dbc-7b5a-4836-91bf-a0aa33cca6bc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.930788] env[61868]: DEBUG nova.compute.manager [req-ff405812-7e2e-433d-8b8e-af7431b78cff req-46f30b3e-001d-438c-986f-fd3ed98dd822 service nova] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Received event network-vif-deleted-6b4cdaf4-10aa-4932-b8ad-9c3d42d2d9cf {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 791.094185] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Releasing lock "refresh_cache-29530178-d69c-4aed-9061-d3d1cfa954d0" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.094429] env[61868]: DEBUG nova.compute.manager [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 791.094594] env[61868]: DEBUG nova.compute.manager [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 791.094754] env[61868]: DEBUG nova.network.neutron [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 791.110786] env[61868]: DEBUG nova.network.neutron [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 791.146025] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Releasing lock "refresh_cache-cd1ccbd7-5d2e-4dd9-a62e-24706759a67a" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.146305] env[61868]: DEBUG nova.compute.manager [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 791.146475] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 791.146998] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-470d1a4d-2cd9-48c5-ac76-791aa4db3af8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.156148] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fd97838-e5e4-4738-adbe-1e98fb8735d0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.180615] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance cd1ccbd7-5d2e-4dd9-a62e-24706759a67a could not be found. [ 791.180862] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 791.181068] env[61868]: INFO nova.compute.manager [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Took 0.03 seconds to destroy the instance on the hypervisor. [ 791.181362] env[61868]: DEBUG oslo.service.loopingcall [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 791.183672] env[61868]: DEBUG nova.compute.manager [-] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 791.183773] env[61868]: DEBUG nova.network.neutron [-] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 791.199322] env[61868]: DEBUG nova.network.neutron [-] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 791.231025] env[61868]: INFO nova.scheduler.client.report [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Deleted allocations for instance ebb09f50-50b8-49bb-b366-5b8906605699 [ 791.284983] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bdfe4f3-4619-48c7-9361-b73c7c93a92c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.292683] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b8fdad1-efe0-4324-8ae7-d8c98e172d2d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.331366] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-700e9134-7b4f-49c6-bb64-473609670272 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.340094] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a869d500-40f2-48d0-98ff-7479666a8303 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.351811] env[61868]: DEBUG nova.compute.provider_tree [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 791.614046] env[61868]: DEBUG nova.network.neutron [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.702131] env[61868]: DEBUG nova.network.neutron [-] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.740953] env[61868]: DEBUG oslo_concurrency.lockutils [None req-02836439-8018-403f-9795-b2ea29efa91a tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "ebb09f50-50b8-49bb-b366-5b8906605699" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 152.320s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.854976] env[61868]: DEBUG nova.scheduler.client.report [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 792.117022] env[61868]: INFO nova.compute.manager [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: 29530178-d69c-4aed-9061-d3d1cfa954d0] Took 1.02 seconds to deallocate network for instance. [ 792.204841] env[61868]: INFO nova.compute.manager [-] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Took 1.02 seconds to deallocate network for instance. [ 792.206988] env[61868]: DEBUG nova.compute.claims [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 792.207183] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.243117] env[61868]: DEBUG nova.compute.manager [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 792.359799] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.405s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.360413] env[61868]: DEBUG nova.compute.manager [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 792.363501] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 15.794s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.363501] env[61868]: DEBUG nova.objects.instance [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61868) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 792.766268] env[61868]: DEBUG oslo_concurrency.lockutils [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.867835] env[61868]: DEBUG nova.compute.utils [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 792.872093] env[61868]: DEBUG nova.compute.manager [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 792.872403] env[61868]: DEBUG nova.network.neutron [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 792.926692] env[61868]: DEBUG nova.policy [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5dee8aa3131d4ad68e93ae8750dafe7e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '25590f7fd009424ca25c03bf16899b5a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 793.154744] env[61868]: INFO nova.scheduler.client.report [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Deleted allocations for instance 29530178-d69c-4aed-9061-d3d1cfa954d0 [ 793.228632] env[61868]: DEBUG nova.network.neutron [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Successfully created port: 3d90f42c-11ae-41f2-ae56-9b45d6a9012c {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 793.373397] env[61868]: DEBUG nova.compute.manager [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 793.377555] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4b071def-bda3-4721-9ae7-817b3946996e tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.378933] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.383s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.666540] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3f948d45-4745-44f0-8b25-da114ade399a tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Lock "29530178-d69c-4aed-9061-d3d1cfa954d0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 149.946s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.716894] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "5aec2609-79d3-4725-a182-70b575adbe17" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.717352] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "5aec2609-79d3-4725-a182-70b575adbe17" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.091067] env[61868]: DEBUG nova.compute.manager [req-6b73faff-3349-4ccc-ae74-dcb3ea0aaec8 req-787e748a-e1ec-4579-a6c2-92024129b66d service nova] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Received event network-changed-3d90f42c-11ae-41f2-ae56-9b45d6a9012c {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 794.091316] env[61868]: DEBUG nova.compute.manager [req-6b73faff-3349-4ccc-ae74-dcb3ea0aaec8 req-787e748a-e1ec-4579-a6c2-92024129b66d service nova] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Refreshing instance network info cache due to event network-changed-3d90f42c-11ae-41f2-ae56-9b45d6a9012c. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 794.091610] env[61868]: DEBUG oslo_concurrency.lockutils [req-6b73faff-3349-4ccc-ae74-dcb3ea0aaec8 req-787e748a-e1ec-4579-a6c2-92024129b66d service nova] Acquiring lock "refresh_cache-5c1683e1-0c87-4d1b-ae21-e07e04d848b1" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.091820] env[61868]: DEBUG oslo_concurrency.lockutils [req-6b73faff-3349-4ccc-ae74-dcb3ea0aaec8 req-787e748a-e1ec-4579-a6c2-92024129b66d service nova] Acquired lock "refresh_cache-5c1683e1-0c87-4d1b-ae21-e07e04d848b1" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.092730] env[61868]: DEBUG nova.network.neutron [req-6b73faff-3349-4ccc-ae74-dcb3ea0aaec8 req-787e748a-e1ec-4579-a6c2-92024129b66d service nova] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Refreshing network info cache for port 3d90f42c-11ae-41f2-ae56-9b45d6a9012c {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 794.102133] env[61868]: ERROR nova.compute.manager [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3d90f42c-11ae-41f2-ae56-9b45d6a9012c, please check neutron logs for more information. [ 794.102133] env[61868]: ERROR nova.compute.manager Traceback (most recent call last): [ 794.102133] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 794.102133] env[61868]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 794.102133] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 794.102133] env[61868]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 794.102133] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 794.102133] env[61868]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 794.102133] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 794.102133] env[61868]: ERROR nova.compute.manager self.force_reraise() [ 794.102133] env[61868]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 794.102133] env[61868]: ERROR nova.compute.manager raise self.value [ 794.102133] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 794.102133] env[61868]: ERROR nova.compute.manager updated_port = self._update_port( [ 794.102133] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 794.102133] env[61868]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 794.102669] env[61868]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 794.102669] env[61868]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 794.102669] env[61868]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3d90f42c-11ae-41f2-ae56-9b45d6a9012c, please check neutron logs for more information. [ 794.102669] env[61868]: ERROR nova.compute.manager [ 794.102669] env[61868]: Traceback (most recent call last): [ 794.102669] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 794.102669] env[61868]: listener.cb(fileno) [ 794.102669] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 794.102669] env[61868]: result = function(*args, **kwargs) [ 794.102669] env[61868]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 794.102669] env[61868]: return func(*args, **kwargs) [ 794.102669] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 794.102669] env[61868]: raise e [ 794.102669] env[61868]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 794.102669] env[61868]: nwinfo = self.network_api.allocate_for_instance( [ 794.102669] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 794.102669] env[61868]: created_port_ids = self._update_ports_for_instance( [ 794.102669] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 794.102669] env[61868]: with excutils.save_and_reraise_exception(): [ 794.102669] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 794.102669] env[61868]: self.force_reraise() [ 794.102669] env[61868]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 794.102669] env[61868]: raise self.value [ 794.102669] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 794.102669] env[61868]: updated_port = self._update_port( [ 794.102669] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 794.102669] env[61868]: _ensure_no_port_binding_failure(port) [ 794.102669] env[61868]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 794.102669] env[61868]: raise exception.PortBindingFailed(port_id=port['id']) [ 794.103562] env[61868]: nova.exception.PortBindingFailed: Binding failed for port 3d90f42c-11ae-41f2-ae56-9b45d6a9012c, please check neutron logs for more information. [ 794.103562] env[61868]: Removing descriptor: 16 [ 794.168332] env[61868]: DEBUG nova.compute.manager [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 794.310967] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f51759aa-9a95-4191-9d66-1a4a0bae7e02 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.318041] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef8f9890-790c-4864-9268-55adcd7407d5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.349097] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85bde61c-cd4a-404d-8d57-b336f19dbda9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.356705] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e782a1a-295f-4091-9180-165d205141ce {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.370172] env[61868]: DEBUG nova.compute.provider_tree [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 794.382672] env[61868]: DEBUG nova.compute.manager [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 794.406083] env[61868]: DEBUG nova.virt.hardware [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 794.406366] env[61868]: DEBUG nova.virt.hardware [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 794.406525] env[61868]: DEBUG nova.virt.hardware [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 794.406704] env[61868]: DEBUG nova.virt.hardware [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 794.406845] env[61868]: DEBUG nova.virt.hardware [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 794.406987] env[61868]: DEBUG nova.virt.hardware [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 794.407199] env[61868]: DEBUG nova.virt.hardware [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 794.407352] env[61868]: DEBUG nova.virt.hardware [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 794.407517] env[61868]: DEBUG nova.virt.hardware [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 794.407697] env[61868]: DEBUG nova.virt.hardware [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 794.407873] env[61868]: DEBUG nova.virt.hardware [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 794.408704] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15b63541-79eb-4810-939e-cc9bdaae4356 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.417090] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a3e27eb-263b-4c4f-8f08-d058a7c2fdd4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.430484] env[61868]: ERROR nova.compute.manager [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3d90f42c-11ae-41f2-ae56-9b45d6a9012c, please check neutron logs for more information. [ 794.430484] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Traceback (most recent call last): [ 794.430484] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 794.430484] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] yield resources [ 794.430484] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 794.430484] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] self.driver.spawn(context, instance, image_meta, [ 794.430484] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 794.430484] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 794.430484] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 794.430484] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] vm_ref = self.build_virtual_machine(instance, [ 794.430484] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 794.430812] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] vif_infos = vmwarevif.get_vif_info(self._session, [ 794.430812] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 794.430812] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] for vif in network_info: [ 794.430812] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 794.430812] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] return self._sync_wrapper(fn, *args, **kwargs) [ 794.430812] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 794.430812] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] self.wait() [ 794.430812] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 794.430812] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] self[:] = self._gt.wait() [ 794.430812] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 794.430812] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] return self._exit_event.wait() [ 794.430812] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 794.430812] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] current.throw(*self._exc) [ 794.431141] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 794.431141] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] result = function(*args, **kwargs) [ 794.431141] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 794.431141] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] return func(*args, **kwargs) [ 794.431141] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 794.431141] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] raise e [ 794.431141] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 794.431141] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] nwinfo = self.network_api.allocate_for_instance( [ 794.431141] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 794.431141] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] created_port_ids = self._update_ports_for_instance( [ 794.431141] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 794.431141] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] with excutils.save_and_reraise_exception(): [ 794.431141] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 794.431509] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] self.force_reraise() [ 794.431509] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 794.431509] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] raise self.value [ 794.431509] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 794.431509] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] updated_port = self._update_port( [ 794.431509] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 794.431509] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] _ensure_no_port_binding_failure(port) [ 794.431509] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 794.431509] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] raise exception.PortBindingFailed(port_id=port['id']) [ 794.431509] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] nova.exception.PortBindingFailed: Binding failed for port 3d90f42c-11ae-41f2-ae56-9b45d6a9012c, please check neutron logs for more information. [ 794.431509] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] [ 794.431509] env[61868]: INFO nova.compute.manager [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Terminating instance [ 794.432814] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] Acquiring lock "refresh_cache-5c1683e1-0c87-4d1b-ae21-e07e04d848b1" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.616559] env[61868]: DEBUG nova.network.neutron [req-6b73faff-3349-4ccc-ae74-dcb3ea0aaec8 req-787e748a-e1ec-4579-a6c2-92024129b66d service nova] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 794.678136] env[61868]: DEBUG nova.network.neutron [req-6b73faff-3349-4ccc-ae74-dcb3ea0aaec8 req-787e748a-e1ec-4579-a6c2-92024129b66d service nova] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.694683] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.876291] env[61868]: DEBUG nova.scheduler.client.report [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 795.183891] env[61868]: DEBUG oslo_concurrency.lockutils [req-6b73faff-3349-4ccc-ae74-dcb3ea0aaec8 req-787e748a-e1ec-4579-a6c2-92024129b66d service nova] Releasing lock "refresh_cache-5c1683e1-0c87-4d1b-ae21-e07e04d848b1" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.184344] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] Acquired lock "refresh_cache-5c1683e1-0c87-4d1b-ae21-e07e04d848b1" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.184528] env[61868]: DEBUG nova.network.neutron [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 795.378733] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.379427] env[61868]: ERROR nova.compute.manager [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 651014ed-1297-4d85-967c-03c06efe3fc2, please check neutron logs for more information. [ 795.379427] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Traceback (most recent call last): [ 795.379427] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 795.379427] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] self.driver.spawn(context, instance, image_meta, [ 795.379427] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 795.379427] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] self._vmops.spawn(context, instance, image_meta, injected_files, [ 795.379427] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 795.379427] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] vm_ref = self.build_virtual_machine(instance, [ 795.379427] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 795.379427] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] vif_infos = vmwarevif.get_vif_info(self._session, [ 795.379427] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 795.379721] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] for vif in network_info: [ 795.379721] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 795.379721] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] return self._sync_wrapper(fn, *args, **kwargs) [ 795.379721] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 795.379721] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] self.wait() [ 795.379721] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 795.379721] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] self[:] = self._gt.wait() [ 795.379721] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 795.379721] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] return self._exit_event.wait() [ 795.379721] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 795.379721] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] current.throw(*self._exc) [ 795.379721] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 795.379721] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] result = function(*args, **kwargs) [ 795.379978] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 795.379978] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] return func(*args, **kwargs) [ 795.379978] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 795.379978] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] raise e [ 795.379978] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 795.379978] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] nwinfo = self.network_api.allocate_for_instance( [ 795.379978] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 795.379978] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] created_port_ids = self._update_ports_for_instance( [ 795.379978] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 795.379978] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] with excutils.save_and_reraise_exception(): [ 795.379978] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 795.379978] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] self.force_reraise() [ 795.379978] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 795.380259] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] raise self.value [ 795.380259] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 795.380259] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] updated_port = self._update_port( [ 795.380259] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 795.380259] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] _ensure_no_port_binding_failure(port) [ 795.380259] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 795.380259] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] raise exception.PortBindingFailed(port_id=port['id']) [ 795.380259] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] nova.exception.PortBindingFailed: Binding failed for port 651014ed-1297-4d85-967c-03c06efe3fc2, please check neutron logs for more information. [ 795.380259] env[61868]: ERROR nova.compute.manager [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] [ 795.380259] env[61868]: DEBUG nova.compute.utils [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Binding failed for port 651014ed-1297-4d85-967c-03c06efe3fc2, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 795.381311] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 15.880s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.384816] env[61868]: DEBUG nova.compute.manager [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Build of instance e023cfb8-5f1c-41d9-9e43-d10e073234ea was re-scheduled: Binding failed for port 651014ed-1297-4d85-967c-03c06efe3fc2, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 795.384816] env[61868]: DEBUG nova.compute.manager [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 795.384816] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] Acquiring lock "refresh_cache-e023cfb8-5f1c-41d9-9e43-d10e073234ea" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.384816] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] Acquired lock "refresh_cache-e023cfb8-5f1c-41d9-9e43-d10e073234ea" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.385099] env[61868]: DEBUG nova.network.neutron [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 795.704506] env[61868]: DEBUG nova.network.neutron [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 795.764048] env[61868]: DEBUG nova.network.neutron [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.908961] env[61868]: DEBUG nova.network.neutron [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 795.984748] env[61868]: DEBUG nova.network.neutron [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.116054] env[61868]: DEBUG nova.compute.manager [req-09391eb8-7188-4991-8589-99dff62e949d req-c00d702e-40fb-4068-b1a6-8542bdf88776 service nova] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Received event network-vif-deleted-3d90f42c-11ae-41f2-ae56-9b45d6a9012c {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 796.266977] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] Releasing lock "refresh_cache-5c1683e1-0c87-4d1b-ae21-e07e04d848b1" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.267258] env[61868]: DEBUG nova.compute.manager [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 796.267404] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 796.267704] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7e069d7c-baba-460f-aa96-01da2a2cddd3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.277008] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dca925b0-59fd-4568-8eef-e330242abab1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.298461] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5c1683e1-0c87-4d1b-ae21-e07e04d848b1 could not be found. [ 796.298732] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 796.298918] env[61868]: INFO nova.compute.manager [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Took 0.03 seconds to destroy the instance on the hypervisor. [ 796.299219] env[61868]: DEBUG oslo.service.loopingcall [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 796.299421] env[61868]: DEBUG nova.compute.manager [-] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 796.299529] env[61868]: DEBUG nova.network.neutron [-] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 796.317286] env[61868]: DEBUG nova.network.neutron [-] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 796.411946] env[61868]: WARNING nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 4de32ee6-e205-435b-a75e-94c3121048e3 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 796.487675] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] Releasing lock "refresh_cache-e023cfb8-5f1c-41d9-9e43-d10e073234ea" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.487909] env[61868]: DEBUG nova.compute.manager [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 796.488127] env[61868]: DEBUG nova.compute.manager [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 796.488298] env[61868]: DEBUG nova.network.neutron [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 796.503615] env[61868]: DEBUG nova.network.neutron [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 796.820063] env[61868]: DEBUG nova.network.neutron [-] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.893068] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Acquiring lock "c6108cf2-b597-4ca7-8f57-12ea82cc6c2c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.893301] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Lock "c6108cf2-b597-4ca7-8f57-12ea82cc6c2c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.916311] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance e023cfb8-5f1c-41d9-9e43-d10e073234ea has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 796.916311] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 6391addb-b1a5-4e15-9bd2-531eb9956b42 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 796.916311] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance c0be3a6b-4d9d-4841-b05c-6aca4662b004 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 796.916311] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance cd1ccbd7-5d2e-4dd9-a62e-24706759a67a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 796.916528] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 5c1683e1-0c87-4d1b-ae21-e07e04d848b1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 797.006339] env[61868]: DEBUG nova.network.neutron [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.325274] env[61868]: INFO nova.compute.manager [-] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Took 1.02 seconds to deallocate network for instance. [ 797.326319] env[61868]: DEBUG nova.compute.claims [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Aborting claim: {{(pid=61868) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 797.326497] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.421037] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 083140bf-233d-49ac-8247-cd93206f9200 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 797.509083] env[61868]: INFO nova.compute.manager [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] [instance: e023cfb8-5f1c-41d9-9e43-d10e073234ea] Took 1.02 seconds to deallocate network for instance. [ 797.923758] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 2c00210e-3b6a-4041-9e8e-7f40ef23d716 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 798.426696] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance d35953d8-692d-498e-baf5-96ef381ce12b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 798.540874] env[61868]: INFO nova.scheduler.client.report [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] Deleted allocations for instance e023cfb8-5f1c-41d9-9e43-d10e073234ea [ 798.930857] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance a16a726b-9673-4de7-9188-19c608a7dc0b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 799.051186] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b6ec0789-c0db-42d0-b1a0-139f80d7c195 tempest-AttachInterfacesUnderV243Test-281329114 tempest-AttachInterfacesUnderV243Test-281329114-project-member] Lock "e023cfb8-5f1c-41d9-9e43-d10e073234ea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 150.612s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.434685] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 799.555634] env[61868]: DEBUG nova.compute.manager [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 799.937236] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 800.150820] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.440273] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 800.944112] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 7788f9f8-c08e-46f8-b204-070215223436 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 801.447139] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 52a5afc6-4e79-436c-bc94-b61ca9fb860c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 801.949917] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance c2693a43-4ea2-4ab0-8915-2fa544780e3c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 802.453507] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 409cad1a-946d-4c58-aa57-1c0bf97fe63f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 802.956651] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance c7b10943-5136-44c8-b4e7-59651220a333 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 803.459748] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance ca85dfd1-e794-41dc-a337-fbd99998e1fa has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 803.963357] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 19c2720c-90bc-47f6-999b-6031f893408d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 804.466105] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 75fdbc11-1ef3-442d-bcf4-55069ba9ab62 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 804.969072] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance c6f99cbc-3e81-44b4-b3d1-d8cd821829f2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 805.471805] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance cf063dbc-7b5a-4836-91bf-a0aa33cca6bc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 805.974550] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 5aec2609-79d3-4725-a182-70b575adbe17 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 805.974821] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=61868) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 805.974998] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=61868) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 806.226989] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83f44d40-5e9b-4978-b670-4f082adc0cbd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.234420] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cff01a9-02e4-4acb-bd67-0be423efb5d5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.262468] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad1e760b-eae3-4967-b040-c12eaa7f4db9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.268969] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37700a86-2136-40f8-973c-b6518d236127 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.281863] env[61868]: DEBUG nova.compute.provider_tree [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 806.786056] env[61868]: DEBUG nova.scheduler.client.report [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 807.291035] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61868) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 807.291035] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 11.909s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.291035] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.556s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.292188] env[61868]: INFO nova.compute.claims [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 807.294935] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 807.295093] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Cleaning up deleted instances {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 807.803767] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] There are 2 instances to clean {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 807.803767] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 7651f6d2-1825-46a1-9380-732b50abd3cc] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 808.308363] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: ddde9ee7-5198-42e5-86b9-f51727dfbf60] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 808.593196] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94665456-08f8-4f7b-be5d-6915b96e6cc6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.600912] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fb6c12a-08ca-4038-b0f3-eda762dd8f35 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.630664] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f52f6a7d-313e-4b1f-8a82-dec421aeee61 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.637832] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6233ace5-fadb-4790-8750-0d3b58e5d9dd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.651484] env[61868]: DEBUG nova.compute.provider_tree [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 808.811926] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 808.812196] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Cleaning up deleted instances with incomplete migration {{(pid=61868) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 809.154851] env[61868]: DEBUG nova.scheduler.client.report [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 809.314807] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 809.659769] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.369s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.660309] env[61868]: DEBUG nova.compute.manager [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 809.662950] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3c4ba62e-0cd1-441f-bc9a-966fa2cf9b11 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.623s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.663151] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3c4ba62e-0cd1-441f-bc9a-966fa2cf9b11 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.665081] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.228s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.689139] env[61868]: INFO nova.scheduler.client.report [None req-3c4ba62e-0cd1-441f-bc9a-966fa2cf9b11 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Deleted allocations for instance 4de32ee6-e205-435b-a75e-94c3121048e3 [ 810.173479] env[61868]: DEBUG nova.compute.utils [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 810.174803] env[61868]: DEBUG nova.compute.manager [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 810.175127] env[61868]: DEBUG nova.network.neutron [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 810.198355] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3c4ba62e-0cd1-441f-bc9a-966fa2cf9b11 tempest-ServerShowV257Test-541215601 tempest-ServerShowV257Test-541215601-project-member] Lock "4de32ee6-e205-435b-a75e-94c3121048e3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.906s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.242718] env[61868]: DEBUG nova.policy [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2b8b2470ec6441f38e7edb0b9ce69020', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f09bf553bf8b47e6846c59523d986edc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 810.556256] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13c85a5d-79ec-4616-8ede-cd2d93461ac8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.566605] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b484aa8-52ae-4033-b0cd-d44ba583dcca {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.605165] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41289b18-6663-4750-b8a8-207ab398e41f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.617050] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed40b210-864b-4a8e-acf3-89db5e800a12 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.622674] env[61868]: DEBUG nova.network.neutron [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Successfully created port: 79fe6cfd-0c91-4ec5-95b2-22aed62a0e0f {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 810.635975] env[61868]: DEBUG nova.compute.provider_tree [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 810.678811] env[61868]: DEBUG nova.compute.manager [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 811.139735] env[61868]: DEBUG nova.scheduler.client.report [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 811.645024] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.979s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.645345] env[61868]: ERROR nova.compute.manager [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4aeee88f-b60a-47e1-bfde-6d563b574608, please check neutron logs for more information. [ 811.645345] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Traceback (most recent call last): [ 811.645345] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 811.645345] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] self.driver.spawn(context, instance, image_meta, [ 811.645345] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 811.645345] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] self._vmops.spawn(context, instance, image_meta, injected_files, [ 811.645345] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 811.645345] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] vm_ref = self.build_virtual_machine(instance, [ 811.645345] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 811.645345] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] vif_infos = vmwarevif.get_vif_info(self._session, [ 811.645345] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 811.645673] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] for vif in network_info: [ 811.645673] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 811.645673] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] return self._sync_wrapper(fn, *args, **kwargs) [ 811.645673] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 811.645673] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] self.wait() [ 811.645673] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 811.645673] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] self[:] = self._gt.wait() [ 811.645673] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 811.645673] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] return self._exit_event.wait() [ 811.645673] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 811.645673] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] current.throw(*self._exc) [ 811.645673] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 811.645673] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] result = function(*args, **kwargs) [ 811.645956] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 811.645956] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] return func(*args, **kwargs) [ 811.645956] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 811.645956] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] raise e [ 811.645956] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 811.645956] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] nwinfo = self.network_api.allocate_for_instance( [ 811.645956] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 811.645956] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] created_port_ids = self._update_ports_for_instance( [ 811.645956] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 811.645956] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] with excutils.save_and_reraise_exception(): [ 811.645956] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 811.645956] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] self.force_reraise() [ 811.645956] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 811.646245] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] raise self.value [ 811.646245] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 811.646245] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] updated_port = self._update_port( [ 811.646245] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 811.646245] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] _ensure_no_port_binding_failure(port) [ 811.646245] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 811.646245] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] raise exception.PortBindingFailed(port_id=port['id']) [ 811.646245] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] nova.exception.PortBindingFailed: Binding failed for port 4aeee88f-b60a-47e1-bfde-6d563b574608, please check neutron logs for more information. [ 811.646245] env[61868]: ERROR nova.compute.manager [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] [ 811.646245] env[61868]: DEBUG nova.compute.utils [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Binding failed for port 4aeee88f-b60a-47e1-bfde-6d563b574608, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 811.647928] env[61868]: DEBUG nova.compute.manager [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Build of instance 6391addb-b1a5-4e15-9bd2-531eb9956b42 was re-scheduled: Binding failed for port 4aeee88f-b60a-47e1-bfde-6d563b574608, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 811.648378] env[61868]: DEBUG nova.compute.manager [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 811.648609] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Acquiring lock "refresh_cache-6391addb-b1a5-4e15-9bd2-531eb9956b42" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.648755] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Acquired lock "refresh_cache-6391addb-b1a5-4e15-9bd2-531eb9956b42" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.648914] env[61868]: DEBUG nova.network.neutron [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 811.649983] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.544s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.653957] env[61868]: INFO nova.compute.claims [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 811.691453] env[61868]: DEBUG nova.compute.manager [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 811.749463] env[61868]: DEBUG nova.virt.hardware [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 811.749463] env[61868]: DEBUG nova.virt.hardware [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 811.749740] env[61868]: DEBUG nova.virt.hardware [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 811.749954] env[61868]: DEBUG nova.virt.hardware [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 811.750153] env[61868]: DEBUG nova.virt.hardware [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 811.750339] env[61868]: DEBUG nova.virt.hardware [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 811.750578] env[61868]: DEBUG nova.virt.hardware [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 811.750765] env[61868]: DEBUG nova.virt.hardware [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 811.750961] env[61868]: DEBUG nova.virt.hardware [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 811.751169] env[61868]: DEBUG nova.virt.hardware [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 811.751445] env[61868]: DEBUG nova.virt.hardware [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 811.752338] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6340d1c-7fec-4cc9-bad3-bdd24579968b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.760888] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec0d4f9a-f872-464d-806e-806b036556fc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.174423] env[61868]: DEBUG nova.network.neutron [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 812.253988] env[61868]: DEBUG nova.network.neutron [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.756181] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Releasing lock "refresh_cache-6391addb-b1a5-4e15-9bd2-531eb9956b42" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.756418] env[61868]: DEBUG nova.compute.manager [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 812.756587] env[61868]: DEBUG nova.compute.manager [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 812.756751] env[61868]: DEBUG nova.network.neutron [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 812.772024] env[61868]: DEBUG nova.network.neutron [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 812.918194] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22d84d35-8079-4b88-bec7-e3299e78c7ef {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.925878] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-505efca2-c363-4b16-b7fb-22696458ab24 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.955387] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88d87f75-2640-495f-b8e2-6e6351da24c2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.962288] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1768c3f8-1195-49bb-adca-36a5fcca0ca3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.974561] env[61868]: DEBUG nova.compute.provider_tree [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 813.275628] env[61868]: DEBUG nova.network.neutron [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.477412] env[61868]: DEBUG nova.scheduler.client.report [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 813.779533] env[61868]: INFO nova.compute.manager [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: 6391addb-b1a5-4e15-9bd2-531eb9956b42] Took 1.02 seconds to deallocate network for instance. [ 813.982906] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.333s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.983319] env[61868]: DEBUG nova.compute.manager [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 813.985927] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.197s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.491746] env[61868]: DEBUG nova.compute.utils [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 814.496179] env[61868]: DEBUG nova.compute.manager [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 814.496269] env[61868]: DEBUG nova.network.neutron [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 814.543720] env[61868]: DEBUG nova.policy [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c780a4ad6d1a4979b6d8580ef5c9c63d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a236c670521a4e1a8d7b047069b07491', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 814.753871] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc513c81-cb49-4e26-8476-1d00d7d023c6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.761584] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e532da7-f205-4fe7-b6c2-43ecc1af63a3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.800036] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e30ee021-fcaa-41c3-a9df-74a89f553f85 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.805754] env[61868]: DEBUG nova.network.neutron [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Successfully created port: bf06b712-731e-430b-bec0-8f026c825d33 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 814.812623] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-506da84a-e812-49ff-b45e-0ccdc9c24075 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.829118] env[61868]: INFO nova.scheduler.client.report [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Deleted allocations for instance 6391addb-b1a5-4e15-9bd2-531eb9956b42 [ 814.834884] env[61868]: DEBUG nova.compute.provider_tree [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 814.997894] env[61868]: DEBUG nova.compute.manager [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 815.339085] env[61868]: DEBUG nova.scheduler.client.report [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 815.344688] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a4b2e01b-e388-409a-97d0-87e40828a9fd tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Lock "6391addb-b1a5-4e15-9bd2-531eb9956b42" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 161.555s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.502796] env[61868]: INFO nova.virt.block_device [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Booting with volume d90113fc-08af-4a33-a1e2-64d3d4549274 at /dev/sda [ 815.535424] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6a021032-7112-4545-be12-7092a3ef991b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.544332] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-805287a1-4309-471b-9b0a-3905baac0096 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.564731] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6282ea4a-4d6f-4aad-a1d1-afe4a6adc932 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.571796] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4902e94-7047-4e94-976a-284e29a2b36f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.592196] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffc6764d-d4d2-41e6-8f23-51f0b9115316 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.598317] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8a91b76-22c6-4d94-b1e6-0c386ef9bc30 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.611173] env[61868]: DEBUG nova.virt.block_device [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Updating existing volume attachment record: b1469c81-8942-47e4-a781-18918a87fcdd {{(pid=61868) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 815.846225] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.860s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.846913] env[61868]: ERROR nova.compute.manager [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c99d18ba-2fcd-4d59-a606-7c99be343788, please check neutron logs for more information. [ 815.846913] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Traceback (most recent call last): [ 815.846913] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 815.846913] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] self.driver.spawn(context, instance, image_meta, [ 815.846913] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 815.846913] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] self._vmops.spawn(context, instance, image_meta, injected_files, [ 815.846913] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 815.846913] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] vm_ref = self.build_virtual_machine(instance, [ 815.846913] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 815.846913] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] vif_infos = vmwarevif.get_vif_info(self._session, [ 815.846913] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 815.847248] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] for vif in network_info: [ 815.847248] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 815.847248] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] return self._sync_wrapper(fn, *args, **kwargs) [ 815.847248] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 815.847248] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] self.wait() [ 815.847248] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 815.847248] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] self[:] = self._gt.wait() [ 815.847248] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 815.847248] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] return self._exit_event.wait() [ 815.847248] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 815.847248] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] current.throw(*self._exc) [ 815.847248] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 815.847248] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] result = function(*args, **kwargs) [ 815.847598] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 815.847598] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] return func(*args, **kwargs) [ 815.847598] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 815.847598] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] raise e [ 815.847598] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 815.847598] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] nwinfo = self.network_api.allocate_for_instance( [ 815.847598] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 815.847598] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] created_port_ids = self._update_ports_for_instance( [ 815.847598] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 815.847598] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] with excutils.save_and_reraise_exception(): [ 815.847598] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 815.847598] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] self.force_reraise() [ 815.847598] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 815.847931] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] raise self.value [ 815.847931] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 815.847931] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] updated_port = self._update_port( [ 815.847931] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 815.847931] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] _ensure_no_port_binding_failure(port) [ 815.847931] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 815.847931] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] raise exception.PortBindingFailed(port_id=port['id']) [ 815.847931] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] nova.exception.PortBindingFailed: Binding failed for port c99d18ba-2fcd-4d59-a606-7c99be343788, please check neutron logs for more information. [ 815.847931] env[61868]: ERROR nova.compute.manager [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] [ 815.847931] env[61868]: DEBUG nova.compute.utils [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Binding failed for port c99d18ba-2fcd-4d59-a606-7c99be343788, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 815.848942] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.315s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.850465] env[61868]: INFO nova.compute.claims [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 815.853028] env[61868]: DEBUG nova.compute.manager [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 815.855599] env[61868]: DEBUG nova.compute.manager [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Build of instance c0be3a6b-4d9d-4841-b05c-6aca4662b004 was re-scheduled: Binding failed for port c99d18ba-2fcd-4d59-a606-7c99be343788, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 815.855744] env[61868]: DEBUG nova.compute.manager [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 815.855885] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "refresh_cache-c0be3a6b-4d9d-4841-b05c-6aca4662b004" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.856035] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquired lock "refresh_cache-c0be3a6b-4d9d-4841-b05c-6aca4662b004" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.856195] env[61868]: DEBUG nova.network.neutron [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 816.357385] env[61868]: DEBUG oslo_concurrency.lockutils [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Acquiring lock "b9b5be37-6b30-4229-9c8d-3ee9d30db119" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.357616] env[61868]: DEBUG oslo_concurrency.lockutils [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Lock "b9b5be37-6b30-4229-9c8d-3ee9d30db119" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.380593] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.385706] env[61868]: DEBUG nova.network.neutron [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 816.390368] env[61868]: DEBUG nova.compute.manager [req-f52dacce-4690-4cd7-bb7a-525a0d2210bc req-a3fa4c52-dbe8-41fe-ad99-cb84b76ae42d service nova] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Received event network-vif-plugged-bf06b712-731e-430b-bec0-8f026c825d33 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 816.390577] env[61868]: DEBUG oslo_concurrency.lockutils [req-f52dacce-4690-4cd7-bb7a-525a0d2210bc req-a3fa4c52-dbe8-41fe-ad99-cb84b76ae42d service nova] Acquiring lock "2c00210e-3b6a-4041-9e8e-7f40ef23d716-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.390776] env[61868]: DEBUG oslo_concurrency.lockutils [req-f52dacce-4690-4cd7-bb7a-525a0d2210bc req-a3fa4c52-dbe8-41fe-ad99-cb84b76ae42d service nova] Lock "2c00210e-3b6a-4041-9e8e-7f40ef23d716-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.390949] env[61868]: DEBUG oslo_concurrency.lockutils [req-f52dacce-4690-4cd7-bb7a-525a0d2210bc req-a3fa4c52-dbe8-41fe-ad99-cb84b76ae42d service nova] Lock "2c00210e-3b6a-4041-9e8e-7f40ef23d716-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.391160] env[61868]: DEBUG nova.compute.manager [req-f52dacce-4690-4cd7-bb7a-525a0d2210bc req-a3fa4c52-dbe8-41fe-ad99-cb84b76ae42d service nova] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] No waiting events found dispatching network-vif-plugged-bf06b712-731e-430b-bec0-8f026c825d33 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 816.391362] env[61868]: WARNING nova.compute.manager [req-f52dacce-4690-4cd7-bb7a-525a0d2210bc req-a3fa4c52-dbe8-41fe-ad99-cb84b76ae42d service nova] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Received unexpected event network-vif-plugged-bf06b712-731e-430b-bec0-8f026c825d33 for instance with vm_state building and task_state block_device_mapping. [ 816.499903] env[61868]: DEBUG nova.network.neutron [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.539747] env[61868]: DEBUG nova.network.neutron [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Successfully updated port: bf06b712-731e-430b-bec0-8f026c825d33 {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 817.003563] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Releasing lock "refresh_cache-c0be3a6b-4d9d-4841-b05c-6aca4662b004" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.003830] env[61868]: DEBUG nova.compute.manager [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 817.004618] env[61868]: DEBUG nova.compute.manager [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 817.004618] env[61868]: DEBUG nova.network.neutron [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 817.019275] env[61868]: DEBUG nova.network.neutron [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 817.042819] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Acquiring lock "refresh_cache-2c00210e-3b6a-4041-9e8e-7f40ef23d716" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.042962] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Acquired lock "refresh_cache-2c00210e-3b6a-4041-9e8e-7f40ef23d716" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.043131] env[61868]: DEBUG nova.network.neutron [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 817.117422] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff21bbdd-1cc7-4653-aefc-b4c00e4f8964 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.126281] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-075ed81c-c56b-4899-a887-abb1f778ad00 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.155430] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91535ce1-fd6e-40fb-9800-4d12bcc6745b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.162806] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a919f13-1d4a-4e5f-ae93-a19619f3a6c4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.175447] env[61868]: DEBUG nova.compute.provider_tree [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 817.522098] env[61868]: DEBUG nova.network.neutron [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.583472] env[61868]: DEBUG nova.network.neutron [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 817.679030] env[61868]: DEBUG nova.scheduler.client.report [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 817.703600] env[61868]: DEBUG nova.compute.manager [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 817.704204] env[61868]: DEBUG nova.virt.hardware [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 817.704454] env[61868]: DEBUG nova.virt.hardware [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 817.704625] env[61868]: DEBUG nova.virt.hardware [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 817.704827] env[61868]: DEBUG nova.virt.hardware [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 817.705048] env[61868]: DEBUG nova.virt.hardware [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 817.705163] env[61868]: DEBUG nova.virt.hardware [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 817.706706] env[61868]: DEBUG nova.virt.hardware [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 817.706706] env[61868]: DEBUG nova.virt.hardware [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 817.706706] env[61868]: DEBUG nova.virt.hardware [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 817.706706] env[61868]: DEBUG nova.virt.hardware [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 817.706706] env[61868]: DEBUG nova.virt.hardware [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 817.706920] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee678a11-9624-4612-bba8-14c412e0cd39 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.716701] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c47bd4e9-60b5-47d8-9ab2-ce12d4e8297b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.764984] env[61868]: DEBUG nova.network.neutron [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Updating instance_info_cache with network_info: [{"id": "bf06b712-731e-430b-bec0-8f026c825d33", "address": "fa:16:3e:1a:e5:fc", "network": {"id": "65e9ddda-7871-4362-b634-eab65fafb929", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-2143875971-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a236c670521a4e1a8d7b047069b07491", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7908211b-df93-467b-87a8-3c3d29b03de6", "external-id": "nsx-vlan-transportzone-632", "segmentation_id": 632, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf06b712-73", "ovs_interfaceid": "bf06b712-731e-430b-bec0-8f026c825d33", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.025185] env[61868]: INFO nova.compute.manager [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: c0be3a6b-4d9d-4841-b05c-6aca4662b004] Took 1.02 seconds to deallocate network for instance. [ 818.184241] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.335s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.184775] env[61868]: DEBUG nova.compute.manager [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 818.187236] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 25.980s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.267438] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Releasing lock "refresh_cache-2c00210e-3b6a-4041-9e8e-7f40ef23d716" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.268091] env[61868]: DEBUG nova.compute.manager [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Instance network_info: |[{"id": "bf06b712-731e-430b-bec0-8f026c825d33", "address": "fa:16:3e:1a:e5:fc", "network": {"id": "65e9ddda-7871-4362-b634-eab65fafb929", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-2143875971-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a236c670521a4e1a8d7b047069b07491", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7908211b-df93-467b-87a8-3c3d29b03de6", "external-id": "nsx-vlan-transportzone-632", "segmentation_id": 632, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf06b712-73", "ovs_interfaceid": "bf06b712-731e-430b-bec0-8f026c825d33", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 818.268811] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1a:e5:fc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7908211b-df93-467b-87a8-3c3d29b03de6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bf06b712-731e-430b-bec0-8f026c825d33', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 818.276895] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Creating folder: Project (a236c670521a4e1a8d7b047069b07491). Parent ref: group-v281478. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 818.277264] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-627d4f4b-e249-4391-ac4b-e9b791e4933c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.292364] env[61868]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 818.292554] env[61868]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=61868) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 818.292852] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Folder already exists: Project (a236c670521a4e1a8d7b047069b07491). Parent ref: group-v281478. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 818.293046] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Creating folder: Instances. Parent ref: group-v281491. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 818.293267] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-af0ee482-7d37-4bbc-bc28-a172d0c7a2e3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.301705] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Created folder: Instances in parent group-v281491. [ 818.301921] env[61868]: DEBUG oslo.service.loopingcall [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 818.302157] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 818.302408] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e48368cc-d18d-459c-a762-b32797396873 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.320912] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 818.320912] env[61868]: value = "task-1315341" [ 818.320912] env[61868]: _type = "Task" [ 818.320912] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.327956] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315341, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.418931] env[61868]: DEBUG nova.compute.manager [req-071abd62-f722-41dc-881c-d54a88ee05bf req-415d37e2-905f-4dda-8988-8051b93d9157 service nova] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Received event network-changed-bf06b712-731e-430b-bec0-8f026c825d33 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 818.419206] env[61868]: DEBUG nova.compute.manager [req-071abd62-f722-41dc-881c-d54a88ee05bf req-415d37e2-905f-4dda-8988-8051b93d9157 service nova] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Refreshing instance network info cache due to event network-changed-bf06b712-731e-430b-bec0-8f026c825d33. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 818.419478] env[61868]: DEBUG oslo_concurrency.lockutils [req-071abd62-f722-41dc-881c-d54a88ee05bf req-415d37e2-905f-4dda-8988-8051b93d9157 service nova] Acquiring lock "refresh_cache-2c00210e-3b6a-4041-9e8e-7f40ef23d716" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.419663] env[61868]: DEBUG oslo_concurrency.lockutils [req-071abd62-f722-41dc-881c-d54a88ee05bf req-415d37e2-905f-4dda-8988-8051b93d9157 service nova] Acquired lock "refresh_cache-2c00210e-3b6a-4041-9e8e-7f40ef23d716" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.419857] env[61868]: DEBUG nova.network.neutron [req-071abd62-f722-41dc-881c-d54a88ee05bf req-415d37e2-905f-4dda-8988-8051b93d9157 service nova] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Refreshing network info cache for port bf06b712-731e-430b-bec0-8f026c825d33 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 818.691641] env[61868]: DEBUG nova.compute.utils [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 818.699012] env[61868]: DEBUG nova.compute.manager [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Not allocating networking since 'none' was specified. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 818.831174] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315341, 'name': CreateVM_Task, 'duration_secs': 0.285546} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.833139] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 818.840301] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'boot_index': 0, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281494', 'volume_id': 'd90113fc-08af-4a33-a1e2-64d3d4549274', 'name': 'volume-d90113fc-08af-4a33-a1e2-64d3d4549274', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2c00210e-3b6a-4041-9e8e-7f40ef23d716', 'attached_at': '', 'detached_at': '', 'volume_id': 'd90113fc-08af-4a33-a1e2-64d3d4549274', 'serial': 'd90113fc-08af-4a33-a1e2-64d3d4549274'}, 'device_type': None, 'delete_on_termination': True, 'disk_bus': None, 'mount_device': '/dev/sda', 'guest_format': None, 'attachment_id': 'b1469c81-8942-47e4-a781-18918a87fcdd', 'volume_type': None}], 'swap': None} {{(pid=61868) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 818.840520] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Root volume attach. Driver type: vmdk {{(pid=61868) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 818.841307] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da59876f-5f02-4bb9-8070-4fcb0133ccdc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.852636] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b064e9d7-817a-4110-a225-76e1c943bc6f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.859587] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75516d46-4e97-4533-a27d-cf584faf49b7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.865414] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-e412bf26-afb9-4799-a8c6-fec3e1c964a8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.873992] env[61868]: DEBUG oslo_vmware.api [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Waiting for the task: (returnval){ [ 818.873992] env[61868]: value = "task-1315342" [ 818.873992] env[61868]: _type = "Task" [ 818.873992] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.885855] env[61868]: DEBUG oslo_vmware.api [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Task: {'id': task-1315342, 'name': RelocateVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.966966] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8462a23-0d44-4a18-98a1-252c510e70f5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.974362] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23a3fd7e-2d03-4e2e-916e-83c3ca2a5575 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.003310] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9703e436-d3b5-4393-8d7a-d8c601831dde {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.011676] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d7de298-c57d-40f9-a4c3-327f299e9102 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.024517] env[61868]: DEBUG nova.compute.provider_tree [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 819.058515] env[61868]: INFO nova.scheduler.client.report [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Deleted allocations for instance c0be3a6b-4d9d-4841-b05c-6aca4662b004 [ 819.164776] env[61868]: DEBUG nova.network.neutron [req-071abd62-f722-41dc-881c-d54a88ee05bf req-415d37e2-905f-4dda-8988-8051b93d9157 service nova] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Updated VIF entry in instance network info cache for port bf06b712-731e-430b-bec0-8f026c825d33. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 819.164776] env[61868]: DEBUG nova.network.neutron [req-071abd62-f722-41dc-881c-d54a88ee05bf req-415d37e2-905f-4dda-8988-8051b93d9157 service nova] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Updating instance_info_cache with network_info: [{"id": "bf06b712-731e-430b-bec0-8f026c825d33", "address": "fa:16:3e:1a:e5:fc", "network": {"id": "65e9ddda-7871-4362-b634-eab65fafb929", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-2143875971-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a236c670521a4e1a8d7b047069b07491", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7908211b-df93-467b-87a8-3c3d29b03de6", "external-id": "nsx-vlan-transportzone-632", "segmentation_id": 632, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf06b712-73", "ovs_interfaceid": "bf06b712-731e-430b-bec0-8f026c825d33", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.199613] env[61868]: DEBUG nova.compute.manager [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 819.384185] env[61868]: DEBUG oslo_vmware.api [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Task: {'id': task-1315342, 'name': RelocateVM_Task, 'duration_secs': 0.359788} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.384465] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Volume attach. Driver type: vmdk {{(pid=61868) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 819.384665] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281494', 'volume_id': 'd90113fc-08af-4a33-a1e2-64d3d4549274', 'name': 'volume-d90113fc-08af-4a33-a1e2-64d3d4549274', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2c00210e-3b6a-4041-9e8e-7f40ef23d716', 'attached_at': '', 'detached_at': '', 'volume_id': 'd90113fc-08af-4a33-a1e2-64d3d4549274', 'serial': 'd90113fc-08af-4a33-a1e2-64d3d4549274'} {{(pid=61868) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 819.385422] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccee82d0-7ab1-44da-849c-9cd4e1628082 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.400424] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffa7749c-c0eb-4da8-ba09-d07d344ac934 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.422100] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Reconfiguring VM instance instance-00000036 to attach disk [datastore1] volume-d90113fc-08af-4a33-a1e2-64d3d4549274/volume-d90113fc-08af-4a33-a1e2-64d3d4549274.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 819.422479] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b195d390-73c1-441c-bd26-5f7a9888c048 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.441808] env[61868]: DEBUG oslo_vmware.api [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Waiting for the task: (returnval){ [ 819.441808] env[61868]: value = "task-1315343" [ 819.441808] env[61868]: _type = "Task" [ 819.441808] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.451268] env[61868]: DEBUG oslo_vmware.api [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Task: {'id': task-1315343, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.529042] env[61868]: DEBUG nova.scheduler.client.report [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 819.566655] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4af7ba80-71c4-4b0a-af60-d1a08480cc99 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "c0be3a6b-4d9d-4841-b05c-6aca4662b004" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 146.007s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.667851] env[61868]: DEBUG oslo_concurrency.lockutils [req-071abd62-f722-41dc-881c-d54a88ee05bf req-415d37e2-905f-4dda-8988-8051b93d9157 service nova] Releasing lock "refresh_cache-2c00210e-3b6a-4041-9e8e-7f40ef23d716" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.951847] env[61868]: DEBUG oslo_vmware.api [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Task: {'id': task-1315343, 'name': ReconfigVM_Task, 'duration_secs': 0.304295} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.952872] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Reconfigured VM instance instance-00000036 to attach disk [datastore1] volume-d90113fc-08af-4a33-a1e2-64d3d4549274/volume-d90113fc-08af-4a33-a1e2-64d3d4549274.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 819.958135] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-52d4c18e-97ec-469a-89cf-bf672b6dccdb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.974025] env[61868]: DEBUG oslo_vmware.api [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Waiting for the task: (returnval){ [ 819.974025] env[61868]: value = "task-1315344" [ 819.974025] env[61868]: _type = "Task" [ 819.974025] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.985257] env[61868]: DEBUG oslo_vmware.api [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Task: {'id': task-1315344, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.034236] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.847s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.034741] env[61868]: ERROR nova.compute.manager [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6b4cdaf4-10aa-4932-b8ad-9c3d42d2d9cf, please check neutron logs for more information. [ 820.034741] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Traceback (most recent call last): [ 820.034741] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 820.034741] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] self.driver.spawn(context, instance, image_meta, [ 820.034741] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 820.034741] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 820.034741] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 820.034741] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] vm_ref = self.build_virtual_machine(instance, [ 820.034741] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 820.034741] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] vif_infos = vmwarevif.get_vif_info(self._session, [ 820.034741] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 820.035058] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] for vif in network_info: [ 820.035058] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 820.035058] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] return self._sync_wrapper(fn, *args, **kwargs) [ 820.035058] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 820.035058] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] self.wait() [ 820.035058] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 820.035058] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] self[:] = self._gt.wait() [ 820.035058] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 820.035058] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] return self._exit_event.wait() [ 820.035058] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 820.035058] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] current.throw(*self._exc) [ 820.035058] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 820.035058] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] result = function(*args, **kwargs) [ 820.035346] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 820.035346] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] return func(*args, **kwargs) [ 820.035346] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 820.035346] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] raise e [ 820.035346] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 820.035346] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] nwinfo = self.network_api.allocate_for_instance( [ 820.035346] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 820.035346] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] created_port_ids = self._update_ports_for_instance( [ 820.035346] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 820.035346] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] with excutils.save_and_reraise_exception(): [ 820.035346] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 820.035346] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] self.force_reraise() [ 820.035346] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 820.035640] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] raise self.value [ 820.035640] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 820.035640] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] updated_port = self._update_port( [ 820.035640] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 820.035640] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] _ensure_no_port_binding_failure(port) [ 820.035640] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 820.035640] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] raise exception.PortBindingFailed(port_id=port['id']) [ 820.035640] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] nova.exception.PortBindingFailed: Binding failed for port 6b4cdaf4-10aa-4932-b8ad-9c3d42d2d9cf, please check neutron logs for more information. [ 820.035640] env[61868]: ERROR nova.compute.manager [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] [ 820.035640] env[61868]: DEBUG nova.compute.utils [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Binding failed for port 6b4cdaf4-10aa-4932-b8ad-9c3d42d2d9cf, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 820.037802] env[61868]: DEBUG oslo_concurrency.lockutils [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.271s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.038580] env[61868]: INFO nova.compute.claims [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 820.041123] env[61868]: DEBUG nova.compute.manager [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Build of instance cd1ccbd7-5d2e-4dd9-a62e-24706759a67a was re-scheduled: Binding failed for port 6b4cdaf4-10aa-4932-b8ad-9c3d42d2d9cf, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 820.041618] env[61868]: DEBUG nova.compute.manager [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 820.041836] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Acquiring lock "refresh_cache-cd1ccbd7-5d2e-4dd9-a62e-24706759a67a" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.041942] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Acquired lock "refresh_cache-cd1ccbd7-5d2e-4dd9-a62e-24706759a67a" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.042117] env[61868]: DEBUG nova.network.neutron [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 820.069402] env[61868]: DEBUG nova.compute.manager [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 820.208609] env[61868]: DEBUG nova.compute.manager [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 820.230378] env[61868]: DEBUG nova.virt.hardware [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 820.230627] env[61868]: DEBUG nova.virt.hardware [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 820.230784] env[61868]: DEBUG nova.virt.hardware [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 820.230962] env[61868]: DEBUG nova.virt.hardware [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 820.231127] env[61868]: DEBUG nova.virt.hardware [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 820.231276] env[61868]: DEBUG nova.virt.hardware [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 820.231538] env[61868]: DEBUG nova.virt.hardware [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 820.231701] env[61868]: DEBUG nova.virt.hardware [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 820.231864] env[61868]: DEBUG nova.virt.hardware [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 820.232036] env[61868]: DEBUG nova.virt.hardware [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 820.232214] env[61868]: DEBUG nova.virt.hardware [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 820.233071] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af12bdda-9bd7-42da-ad79-bc300166dbe0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.241971] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cc6dbd4-6d9c-4e0f-a6f3-2c51e52745bc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.997026] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Instance VIF info [] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 821.001341] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Creating folder: Project (79aa437dc0aa42478480028a1686c2e7). Parent ref: group-v281478. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 821.011235] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-44ecdc32-0893-40dc-a924-e283f71ddfbf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.013305] env[61868]: DEBUG oslo_concurrency.lockutils [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "b7055e5a-37d2-42d5-bab0-1e70faf52b58" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.013516] env[61868]: DEBUG oslo_concurrency.lockutils [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "b7055e5a-37d2-42d5-bab0-1e70faf52b58" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.020287] env[61868]: DEBUG oslo_vmware.api [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Task: {'id': task-1315344, 'name': ReconfigVM_Task, 'duration_secs': 0.125668} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.020771] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281494', 'volume_id': 'd90113fc-08af-4a33-a1e2-64d3d4549274', 'name': 'volume-d90113fc-08af-4a33-a1e2-64d3d4549274', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2c00210e-3b6a-4041-9e8e-7f40ef23d716', 'attached_at': '', 'detached_at': '', 'volume_id': 'd90113fc-08af-4a33-a1e2-64d3d4549274', 'serial': 'd90113fc-08af-4a33-a1e2-64d3d4549274'} {{(pid=61868) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 821.021279] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b5c1f9f3-996f-456a-8e09-33c36b14ee6e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.023935] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Created folder: Project (79aa437dc0aa42478480028a1686c2e7) in parent group-v281478. [ 821.024124] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Creating folder: Instances. Parent ref: group-v281501. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 821.024666] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9157df06-ea00-4d6b-b572-db49a56fc4f4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.028051] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.029936] env[61868]: DEBUG nova.network.neutron [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 821.031825] env[61868]: DEBUG oslo_vmware.api [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Waiting for the task: (returnval){ [ 821.031825] env[61868]: value = "task-1315346" [ 821.031825] env[61868]: _type = "Task" [ 821.031825] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.038072] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Created folder: Instances in parent group-v281501. [ 821.038664] env[61868]: DEBUG oslo.service.loopingcall [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 821.038854] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 821.039352] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8f1b724c-20a6-4b02-8e2b-1d23ddc408f0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.057019] env[61868]: DEBUG oslo_vmware.api [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Task: {'id': task-1315346, 'name': Rename_Task} progress is 10%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.062189] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 821.062189] env[61868]: value = "task-1315348" [ 821.062189] env[61868]: _type = "Task" [ 821.062189] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.074258] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315348, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.114419] env[61868]: DEBUG nova.network.neutron [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.387604] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56cb626b-67e3-4c11-9e2b-b558c0ced26c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.395307] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a231828-a53e-47b8-88da-464dcbae3f33 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.425049] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-732fd2a9-1a73-49ef-ab93-cb85ca3d2aba {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.431793] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6e29bb6-496d-4c2d-be5e-d535af25d405 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.444435] env[61868]: DEBUG nova.compute.provider_tree [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Updating inventory in ProviderTree for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 821.543736] env[61868]: DEBUG oslo_vmware.api [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Task: {'id': task-1315346, 'name': Rename_Task, 'duration_secs': 0.149947} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.544026] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 821.544247] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2f9c7725-198f-45d3-99c5-548b4f48194d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.550586] env[61868]: DEBUG oslo_vmware.api [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Waiting for the task: (returnval){ [ 821.550586] env[61868]: value = "task-1315349" [ 821.550586] env[61868]: _type = "Task" [ 821.550586] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.557515] env[61868]: DEBUG oslo_vmware.api [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Task: {'id': task-1315349, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.569313] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315348, 'name': CreateVM_Task, 'duration_secs': 0.25376} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.569463] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 821.569834] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.569984] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.570335] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 821.570545] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-118905a4-53bf-475a-85c7-e5ac925b9a2d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.574227] env[61868]: DEBUG oslo_vmware.api [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Waiting for the task: (returnval){ [ 821.574227] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52c70e21-9f09-ab1a-5cd4-e451c418ea57" [ 821.574227] env[61868]: _type = "Task" [ 821.574227] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.580990] env[61868]: DEBUG oslo_vmware.api [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52c70e21-9f09-ab1a-5cd4-e451c418ea57, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.619990] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Releasing lock "refresh_cache-cd1ccbd7-5d2e-4dd9-a62e-24706759a67a" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.620246] env[61868]: DEBUG nova.compute.manager [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 821.620427] env[61868]: DEBUG nova.compute.manager [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 821.620592] env[61868]: DEBUG nova.network.neutron [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 821.636571] env[61868]: DEBUG nova.network.neutron [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 821.968069] env[61868]: ERROR nova.scheduler.client.report [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [req-30f73529-299f-4600-a768-63072877bc81] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 83e67721-2ac9-4a23-aa31-82aca86979c8. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-30f73529-299f-4600-a768-63072877bc81"}]} [ 821.985508] env[61868]: DEBUG nova.scheduler.client.report [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Refreshing inventories for resource provider 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 822.001832] env[61868]: DEBUG nova.scheduler.client.report [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Updating ProviderTree inventory for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 822.001832] env[61868]: DEBUG nova.compute.provider_tree [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Updating inventory in ProviderTree for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 822.014529] env[61868]: DEBUG nova.scheduler.client.report [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Refreshing aggregate associations for resource provider 83e67721-2ac9-4a23-aa31-82aca86979c8, aggregates: None {{(pid=61868) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 822.032517] env[61868]: DEBUG nova.scheduler.client.report [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Refreshing trait associations for resource provider 83e67721-2ac9-4a23-aa31-82aca86979c8, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61868) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 822.059530] env[61868]: DEBUG oslo_vmware.api [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Task: {'id': task-1315349, 'name': PowerOnVM_Task, 'duration_secs': 0.437008} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.061810] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 822.062026] env[61868]: INFO nova.compute.manager [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Took 4.36 seconds to spawn the instance on the hypervisor. [ 822.062205] env[61868]: DEBUG nova.compute.manager [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 822.063337] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f010ed8-6212-4a53-93ea-bc8691c08f66 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.083208] env[61868]: DEBUG oslo_vmware.api [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52c70e21-9f09-ab1a-5cd4-e451c418ea57, 'name': SearchDatastore_Task, 'duration_secs': 0.010978} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.085508] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.085738] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 822.086685] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.086685] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.086685] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 822.086933] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f1a992a0-3577-48c5-a8fc-55c7a881967b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.094931] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 822.095076] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 822.097934] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d573fbad-5f6a-4d0b-b6f4-ea09dfc0bc2c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.103453] env[61868]: DEBUG oslo_vmware.api [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Waiting for the task: (returnval){ [ 822.103453] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52addfe9-9714-9193-18f0-ae60fe105f14" [ 822.103453] env[61868]: _type = "Task" [ 822.103453] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.111867] env[61868]: DEBUG oslo_vmware.api [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52addfe9-9714-9193-18f0-ae60fe105f14, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.140058] env[61868]: DEBUG nova.network.neutron [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.317996] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74e17439-2a0e-4b2c-831e-8521054ca884 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.325795] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7f1c440-4d3c-437e-9a16-c32175326544 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.355149] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c0d9c39-2a81-4831-926e-139e2bebb65a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.361583] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8e34da8-6c10-43b0-8f2b-92bc1ddcfcc1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.374145] env[61868]: DEBUG nova.compute.provider_tree [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Updating inventory in ProviderTree for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 822.581707] env[61868]: INFO nova.compute.manager [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Took 38.49 seconds to build instance. [ 822.613451] env[61868]: DEBUG oslo_vmware.api [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52addfe9-9714-9193-18f0-ae60fe105f14, 'name': SearchDatastore_Task, 'duration_secs': 0.008661} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.614178] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b485381d-5a55-4c12-a591-5a03bcfc9e83 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.618884] env[61868]: DEBUG oslo_vmware.api [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Waiting for the task: (returnval){ [ 822.618884] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52aeac3a-6110-6305-d157-0de05702d184" [ 822.618884] env[61868]: _type = "Task" [ 822.618884] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.626361] env[61868]: DEBUG oslo_vmware.api [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52aeac3a-6110-6305-d157-0de05702d184, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.642216] env[61868]: INFO nova.compute.manager [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] [instance: cd1ccbd7-5d2e-4dd9-a62e-24706759a67a] Took 1.02 seconds to deallocate network for instance. [ 822.903961] env[61868]: DEBUG nova.scheduler.client.report [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Updated inventory for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 with generation 84 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 822.904309] env[61868]: DEBUG nova.compute.provider_tree [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Updating resource provider 83e67721-2ac9-4a23-aa31-82aca86979c8 generation from 84 to 85 during operation: update_inventory {{(pid=61868) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 822.904491] env[61868]: DEBUG nova.compute.provider_tree [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Updating inventory in ProviderTree for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 135, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 823.083450] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0122e58e-4f8b-4381-a9ee-fa74d46a5e6a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Lock "2c00210e-3b6a-4041-9e8e-7f40ef23d716" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 111.643s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.130247] env[61868]: DEBUG oslo_vmware.api [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52aeac3a-6110-6305-d157-0de05702d184, 'name': SearchDatastore_Task, 'duration_secs': 0.00893} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.130493] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.130744] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] d35953d8-692d-498e-baf5-96ef381ce12b/d35953d8-692d-498e-baf5-96ef381ce12b.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 823.130987] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ca432132-dce9-47dd-ae41-417228e67ca5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.137707] env[61868]: DEBUG oslo_vmware.api [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Waiting for the task: (returnval){ [ 823.137707] env[61868]: value = "task-1315350" [ 823.137707] env[61868]: _type = "Task" [ 823.137707] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.148416] env[61868]: DEBUG oslo_vmware.api [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315350, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.409488] env[61868]: DEBUG oslo_concurrency.lockutils [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.372s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.410065] env[61868]: DEBUG nova.compute.manager [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 823.413517] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.719s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.415120] env[61868]: INFO nova.compute.claims [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 823.586216] env[61868]: DEBUG nova.compute.manager [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 823.647664] env[61868]: DEBUG oslo_vmware.api [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315350, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.672942] env[61868]: INFO nova.scheduler.client.report [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Deleted allocations for instance cd1ccbd7-5d2e-4dd9-a62e-24706759a67a [ 823.922397] env[61868]: DEBUG nova.compute.utils [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 823.923893] env[61868]: DEBUG nova.compute.manager [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Not allocating networking since 'none' was specified. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 823.953777] env[61868]: DEBUG nova.compute.manager [req-40c3df9e-f514-4314-ac1a-93b0f6155ec1 req-eec0344c-d2a2-4de2-8f5b-5e83e0161dd0 service nova] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Received event network-changed-bf06b712-731e-430b-bec0-8f026c825d33 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 823.953777] env[61868]: DEBUG nova.compute.manager [req-40c3df9e-f514-4314-ac1a-93b0f6155ec1 req-eec0344c-d2a2-4de2-8f5b-5e83e0161dd0 service nova] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Refreshing instance network info cache due to event network-changed-bf06b712-731e-430b-bec0-8f026c825d33. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 823.954103] env[61868]: DEBUG oslo_concurrency.lockutils [req-40c3df9e-f514-4314-ac1a-93b0f6155ec1 req-eec0344c-d2a2-4de2-8f5b-5e83e0161dd0 service nova] Acquiring lock "refresh_cache-2c00210e-3b6a-4041-9e8e-7f40ef23d716" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.954322] env[61868]: DEBUG oslo_concurrency.lockutils [req-40c3df9e-f514-4314-ac1a-93b0f6155ec1 req-eec0344c-d2a2-4de2-8f5b-5e83e0161dd0 service nova] Acquired lock "refresh_cache-2c00210e-3b6a-4041-9e8e-7f40ef23d716" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.954572] env[61868]: DEBUG nova.network.neutron [req-40c3df9e-f514-4314-ac1a-93b0f6155ec1 req-eec0344c-d2a2-4de2-8f5b-5e83e0161dd0 service nova] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Refreshing network info cache for port bf06b712-731e-430b-bec0-8f026c825d33 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 824.108611] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.149834] env[61868]: DEBUG oslo_vmware.api [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315350, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.510675} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.149834] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] d35953d8-692d-498e-baf5-96ef381ce12b/d35953d8-692d-498e-baf5-96ef381ce12b.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 824.149834] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 824.149834] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-614a2a7e-dcc9-4359-bb2c-f59fd7af6211 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.155648] env[61868]: DEBUG oslo_vmware.api [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Waiting for the task: (returnval){ [ 824.155648] env[61868]: value = "task-1315351" [ 824.155648] env[61868]: _type = "Task" [ 824.155648] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.163455] env[61868]: DEBUG oslo_vmware.api [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315351, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.181121] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0d4d4ce8-de23-4777-9771-42b60b50b61f tempest-ServerGroupTestJSON-350065116 tempest-ServerGroupTestJSON-350065116-project-member] Lock "cd1ccbd7-5d2e-4dd9-a62e-24706759a67a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 138.770s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.428137] env[61868]: DEBUG nova.compute.manager [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 824.668916] env[61868]: DEBUG oslo_vmware.api [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315351, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063659} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.669111] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 824.669976] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a27e564-15db-430e-802e-d0fb80a9b961 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.686133] env[61868]: DEBUG nova.compute.manager [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 824.704063] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Reconfiguring VM instance instance-00000037 to attach disk [datastore1] d35953d8-692d-498e-baf5-96ef381ce12b/d35953d8-692d-498e-baf5-96ef381ce12b.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 824.707208] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aa8f035b-7274-4c38-ba03-460145902a83 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.724827] env[61868]: DEBUG nova.network.neutron [req-40c3df9e-f514-4314-ac1a-93b0f6155ec1 req-eec0344c-d2a2-4de2-8f5b-5e83e0161dd0 service nova] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Updated VIF entry in instance network info cache for port bf06b712-731e-430b-bec0-8f026c825d33. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 824.725247] env[61868]: DEBUG nova.network.neutron [req-40c3df9e-f514-4314-ac1a-93b0f6155ec1 req-eec0344c-d2a2-4de2-8f5b-5e83e0161dd0 service nova] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Updating instance_info_cache with network_info: [{"id": "bf06b712-731e-430b-bec0-8f026c825d33", "address": "fa:16:3e:1a:e5:fc", "network": {"id": "65e9ddda-7871-4362-b634-eab65fafb929", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-2143875971-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.221", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a236c670521a4e1a8d7b047069b07491", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7908211b-df93-467b-87a8-3c3d29b03de6", "external-id": "nsx-vlan-transportzone-632", "segmentation_id": 632, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf06b712-73", "ovs_interfaceid": "bf06b712-731e-430b-bec0-8f026c825d33", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.732471] env[61868]: DEBUG oslo_vmware.api [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Waiting for the task: (returnval){ [ 824.732471] env[61868]: value = "task-1315352" [ 824.732471] env[61868]: _type = "Task" [ 824.732471] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.742389] env[61868]: DEBUG oslo_vmware.api [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315352, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.761173] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de6c510-ffa6-42e7-89cc-fadde5854cbe {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.768116] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8070ce98-f4dd-4873-8291-e2ab9b36f80d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.797438] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b3c0150-949d-45bb-a644-9a5cf625ec50 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.804581] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94f12594-e6f6-4cd6-b08f-469b0ae897a7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.817433] env[61868]: DEBUG nova.compute.provider_tree [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Updating inventory in ProviderTree for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 825.224227] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.228148] env[61868]: DEBUG oslo_concurrency.lockutils [req-40c3df9e-f514-4314-ac1a-93b0f6155ec1 req-eec0344c-d2a2-4de2-8f5b-5e83e0161dd0 service nova] Releasing lock "refresh_cache-2c00210e-3b6a-4041-9e8e-7f40ef23d716" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.246908] env[61868]: DEBUG oslo_vmware.api [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315352, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.356139] env[61868]: DEBUG nova.scheduler.client.report [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Updated inventory for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 with generation 85 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 825.356139] env[61868]: DEBUG nova.compute.provider_tree [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Updating resource provider 83e67721-2ac9-4a23-aa31-82aca86979c8 generation from 85 to 86 during operation: update_inventory {{(pid=61868) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 825.356372] env[61868]: DEBUG nova.compute.provider_tree [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Updating inventory in ProviderTree for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 825.438955] env[61868]: DEBUG nova.compute.manager [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 825.471657] env[61868]: DEBUG nova.virt.hardware [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 825.471941] env[61868]: DEBUG nova.virt.hardware [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 825.472115] env[61868]: DEBUG nova.virt.hardware [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 825.472295] env[61868]: DEBUG nova.virt.hardware [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 825.472540] env[61868]: DEBUG nova.virt.hardware [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 825.472634] env[61868]: DEBUG nova.virt.hardware [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 825.472772] env[61868]: DEBUG nova.virt.hardware [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 825.472928] env[61868]: DEBUG nova.virt.hardware [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 825.473113] env[61868]: DEBUG nova.virt.hardware [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 825.473275] env[61868]: DEBUG nova.virt.hardware [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 825.473566] env[61868]: DEBUG nova.virt.hardware [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 825.474475] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14d1c970-5af3-4b53-9116-3600deb7445a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.482872] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49608a77-438f-475a-9b0f-b6f883108963 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.496555] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Instance VIF info [] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 825.501769] env[61868]: DEBUG oslo.service.loopingcall [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 825.501962] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 825.502189] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-32c63293-1ef7-48cb-ac14-943d47097733 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.519854] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 825.519854] env[61868]: value = "task-1315353" [ 825.519854] env[61868]: _type = "Task" [ 825.519854] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.528239] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315353, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.742916] env[61868]: DEBUG oslo_vmware.api [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315352, 'name': ReconfigVM_Task, 'duration_secs': 0.804685} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.743222] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Reconfigured VM instance instance-00000037 to attach disk [datastore1] d35953d8-692d-498e-baf5-96ef381ce12b/d35953d8-692d-498e-baf5-96ef381ce12b.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 825.743809] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f836bf67-ac26-4f51-935b-0baf092ae0e0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.749669] env[61868]: DEBUG oslo_vmware.api [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Waiting for the task: (returnval){ [ 825.749669] env[61868]: value = "task-1315354" [ 825.749669] env[61868]: _type = "Task" [ 825.749669] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.762399] env[61868]: DEBUG oslo_vmware.api [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315354, 'name': Rename_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.868256] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.455s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.868842] env[61868]: DEBUG nova.compute.manager [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 825.872474] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.546s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.030270] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315353, 'name': CreateVM_Task, 'duration_secs': 0.386942} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.030670] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 826.031218] env[61868]: DEBUG oslo_concurrency.lockutils [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.031492] env[61868]: DEBUG oslo_concurrency.lockutils [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.031887] env[61868]: DEBUG oslo_concurrency.lockutils [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 826.032261] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eddc033a-ff9e-4254-ae51-f16b238f1601 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.036872] env[61868]: DEBUG oslo_vmware.api [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Waiting for the task: (returnval){ [ 826.036872] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5223f58e-0f7b-de85-ed9e-5476bc8b49bc" [ 826.036872] env[61868]: _type = "Task" [ 826.036872] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.044589] env[61868]: DEBUG oslo_vmware.api [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5223f58e-0f7b-de85-ed9e-5476bc8b49bc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.259929] env[61868]: DEBUG oslo_vmware.api [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315354, 'name': Rename_Task, 'duration_secs': 0.14983} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.260233] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 826.260785] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-233b0a89-4db5-4243-ba54-43719367e1bc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.266736] env[61868]: DEBUG oslo_vmware.api [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Waiting for the task: (returnval){ [ 826.266736] env[61868]: value = "task-1315355" [ 826.266736] env[61868]: _type = "Task" [ 826.266736] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.275587] env[61868]: DEBUG oslo_vmware.api [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315355, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.382678] env[61868]: DEBUG nova.compute.utils [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 826.387181] env[61868]: DEBUG nova.compute.manager [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 826.387482] env[61868]: DEBUG nova.network.neutron [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 826.444852] env[61868]: DEBUG nova.policy [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fd6ed09d33f6491b83dc72cef31bfd29', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '812e115cc1114d0e8536924c4f5556f8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 826.547940] env[61868]: DEBUG oslo_vmware.api [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5223f58e-0f7b-de85-ed9e-5476bc8b49bc, 'name': SearchDatastore_Task, 'duration_secs': 0.022808} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.548240] env[61868]: DEBUG oslo_concurrency.lockutils [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.548725] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 826.549683] env[61868]: DEBUG oslo_concurrency.lockutils [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.549683] env[61868]: DEBUG oslo_concurrency.lockutils [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.549683] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 826.549683] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-20be382a-ee8e-4a41-851a-93b44a36de1e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.559696] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 826.559871] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 826.560626] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd422e60-ed96-403e-a9fe-0f23ea270e69 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.566865] env[61868]: DEBUG oslo_vmware.api [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Waiting for the task: (returnval){ [ 826.566865] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52c29a65-7a82-eb2d-64ae-5d73535c10ad" [ 826.566865] env[61868]: _type = "Task" [ 826.566865] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.574094] env[61868]: DEBUG oslo_vmware.api [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52c29a65-7a82-eb2d-64ae-5d73535c10ad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.691461] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-296e77ec-b11c-4fdb-803e-9925782ad7e4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.702050] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-711be02d-3e4e-44ed-88f4-9087ba718cab {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.729801] env[61868]: DEBUG nova.network.neutron [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Successfully created port: 548fe1c2-07a5-45a2-b084-cc95216cdcad {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 826.732804] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-977639d8-052f-488b-9c0e-f5e4e9389c35 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.739412] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52f397c8-2e8b-47ce-9010-8dea41e9a54f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.752709] env[61868]: DEBUG nova.compute.provider_tree [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 826.775831] env[61868]: DEBUG oslo_vmware.api [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315355, 'name': PowerOnVM_Task, 'duration_secs': 0.473397} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.776108] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 826.776306] env[61868]: INFO nova.compute.manager [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Took 6.57 seconds to spawn the instance on the hypervisor. [ 826.776481] env[61868]: DEBUG nova.compute.manager [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 826.777225] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acdacb3e-13b1-43d1-bb0d-eb42a9dd65e3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.888286] env[61868]: DEBUG nova.compute.manager [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 827.077916] env[61868]: DEBUG oslo_vmware.api [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52c29a65-7a82-eb2d-64ae-5d73535c10ad, 'name': SearchDatastore_Task, 'duration_secs': 0.008482} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.079027] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ebcc8841-e192-40bf-a95c-6aed1ea982ec {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.084567] env[61868]: DEBUG oslo_vmware.api [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Waiting for the task: (returnval){ [ 827.084567] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]523b0bfb-aad2-78e1-de57-058d2f49f78b" [ 827.084567] env[61868]: _type = "Task" [ 827.084567] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.092219] env[61868]: DEBUG oslo_vmware.api [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]523b0bfb-aad2-78e1-de57-058d2f49f78b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.256646] env[61868]: DEBUG nova.scheduler.client.report [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 827.293399] env[61868]: INFO nova.compute.manager [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Took 38.78 seconds to build instance. [ 827.598572] env[61868]: DEBUG oslo_vmware.api [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]523b0bfb-aad2-78e1-de57-058d2f49f78b, 'name': SearchDatastore_Task, 'duration_secs': 0.043194} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.598572] env[61868]: DEBUG oslo_concurrency.lockutils [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.598572] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02/02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 827.598572] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6a1ccf65-d4e7-45c8-a7c8-624d63751e38 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.605935] env[61868]: DEBUG oslo_vmware.api [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Waiting for the task: (returnval){ [ 827.605935] env[61868]: value = "task-1315356" [ 827.605935] env[61868]: _type = "Task" [ 827.605935] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.614071] env[61868]: DEBUG oslo_vmware.api [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315356, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.762806] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.890s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.762897] env[61868]: ERROR nova.compute.manager [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3d90f42c-11ae-41f2-ae56-9b45d6a9012c, please check neutron logs for more information. [ 827.762897] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Traceback (most recent call last): [ 827.762897] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 827.762897] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] self.driver.spawn(context, instance, image_meta, [ 827.762897] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 827.762897] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 827.762897] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 827.762897] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] vm_ref = self.build_virtual_machine(instance, [ 827.762897] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 827.762897] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] vif_infos = vmwarevif.get_vif_info(self._session, [ 827.762897] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 827.763258] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] for vif in network_info: [ 827.763258] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 827.763258] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] return self._sync_wrapper(fn, *args, **kwargs) [ 827.763258] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 827.763258] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] self.wait() [ 827.763258] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 827.763258] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] self[:] = self._gt.wait() [ 827.763258] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 827.763258] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] return self._exit_event.wait() [ 827.763258] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 827.763258] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] current.throw(*self._exc) [ 827.763258] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 827.763258] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] result = function(*args, **kwargs) [ 827.763569] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 827.763569] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] return func(*args, **kwargs) [ 827.763569] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 827.763569] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] raise e [ 827.763569] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 827.763569] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] nwinfo = self.network_api.allocate_for_instance( [ 827.763569] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 827.763569] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] created_port_ids = self._update_ports_for_instance( [ 827.763569] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 827.763569] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] with excutils.save_and_reraise_exception(): [ 827.763569] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 827.763569] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] self.force_reraise() [ 827.763569] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 827.763891] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] raise self.value [ 827.763891] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 827.763891] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] updated_port = self._update_port( [ 827.763891] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 827.763891] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] _ensure_no_port_binding_failure(port) [ 827.763891] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 827.763891] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] raise exception.PortBindingFailed(port_id=port['id']) [ 827.763891] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] nova.exception.PortBindingFailed: Binding failed for port 3d90f42c-11ae-41f2-ae56-9b45d6a9012c, please check neutron logs for more information. [ 827.763891] env[61868]: ERROR nova.compute.manager [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] [ 827.763891] env[61868]: DEBUG nova.compute.utils [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Binding failed for port 3d90f42c-11ae-41f2-ae56-9b45d6a9012c, please check neutron logs for more information. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 827.764951] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.614s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.766528] env[61868]: INFO nova.compute.claims [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 827.769556] env[61868]: DEBUG nova.compute.manager [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Build of instance 5c1683e1-0c87-4d1b-ae21-e07e04d848b1 was re-scheduled: Binding failed for port 3d90f42c-11ae-41f2-ae56-9b45d6a9012c, please check neutron logs for more information. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 827.770402] env[61868]: DEBUG nova.compute.manager [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 827.770402] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] Acquiring lock "refresh_cache-5c1683e1-0c87-4d1b-ae21-e07e04d848b1" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.770529] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] Acquired lock "refresh_cache-5c1683e1-0c87-4d1b-ae21-e07e04d848b1" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.770649] env[61868]: DEBUG nova.network.neutron [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 827.794315] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0531c8f4-54fc-4fca-913a-49e8a5734bae tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Lock "d35953d8-692d-498e-baf5-96ef381ce12b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 114.683s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.897363] env[61868]: DEBUG nova.compute.manager [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 827.931485] env[61868]: DEBUG nova.virt.hardware [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 827.931788] env[61868]: DEBUG nova.virt.hardware [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 827.931991] env[61868]: DEBUG nova.virt.hardware [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 827.932388] env[61868]: DEBUG nova.virt.hardware [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 827.932472] env[61868]: DEBUG nova.virt.hardware [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 827.932679] env[61868]: DEBUG nova.virt.hardware [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 827.932983] env[61868]: DEBUG nova.virt.hardware [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 827.933180] env[61868]: DEBUG nova.virt.hardware [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 827.933429] env[61868]: DEBUG nova.virt.hardware [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 827.933803] env[61868]: DEBUG nova.virt.hardware [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 827.933944] env[61868]: DEBUG nova.virt.hardware [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 827.935025] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a9d5b3c-04ac-4b47-9af6-a4afdb35ba28 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.944614] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86566e27-8134-4f30-8fdf-a15f32af7f37 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.114979] env[61868]: DEBUG oslo_vmware.api [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315356, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.299469] env[61868]: DEBUG nova.compute.manager [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 828.309496] env[61868]: DEBUG nova.network.neutron [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 828.364925] env[61868]: DEBUG nova.compute.manager [req-6874a376-296c-4afc-b057-a1c6f4527441 req-c926df3d-cc41-4b16-904c-f6a0155821d2 service nova] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Received event network-vif-plugged-548fe1c2-07a5-45a2-b084-cc95216cdcad {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 828.364925] env[61868]: DEBUG oslo_concurrency.lockutils [req-6874a376-296c-4afc-b057-a1c6f4527441 req-c926df3d-cc41-4b16-904c-f6a0155821d2 service nova] Acquiring lock "a16a726b-9673-4de7-9188-19c608a7dc0b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.364925] env[61868]: DEBUG oslo_concurrency.lockutils [req-6874a376-296c-4afc-b057-a1c6f4527441 req-c926df3d-cc41-4b16-904c-f6a0155821d2 service nova] Lock "a16a726b-9673-4de7-9188-19c608a7dc0b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.364925] env[61868]: DEBUG oslo_concurrency.lockutils [req-6874a376-296c-4afc-b057-a1c6f4527441 req-c926df3d-cc41-4b16-904c-f6a0155821d2 service nova] Lock "a16a726b-9673-4de7-9188-19c608a7dc0b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.364925] env[61868]: DEBUG nova.compute.manager [req-6874a376-296c-4afc-b057-a1c6f4527441 req-c926df3d-cc41-4b16-904c-f6a0155821d2 service nova] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] No waiting events found dispatching network-vif-plugged-548fe1c2-07a5-45a2-b084-cc95216cdcad {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 828.365231] env[61868]: WARNING nova.compute.manager [req-6874a376-296c-4afc-b057-a1c6f4527441 req-c926df3d-cc41-4b16-904c-f6a0155821d2 service nova] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Received unexpected event network-vif-plugged-548fe1c2-07a5-45a2-b084-cc95216cdcad for instance with vm_state building and task_state spawning. [ 828.426131] env[61868]: DEBUG nova.network.neutron [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.532472] env[61868]: DEBUG nova.network.neutron [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Successfully updated port: 548fe1c2-07a5-45a2-b084-cc95216cdcad {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 828.618571] env[61868]: DEBUG oslo_vmware.api [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315356, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.522625} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.618837] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02/02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 828.619059] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 828.619321] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1aca94de-fb67-4861-b100-6f56ec8e0d7e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.629354] env[61868]: DEBUG oslo_vmware.api [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Waiting for the task: (returnval){ [ 828.629354] env[61868]: value = "task-1315357" [ 828.629354] env[61868]: _type = "Task" [ 828.629354] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.637556] env[61868]: DEBUG oslo_vmware.api [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315357, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.818037] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.930565] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] Releasing lock "refresh_cache-5c1683e1-0c87-4d1b-ae21-e07e04d848b1" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.930565] env[61868]: DEBUG nova.compute.manager [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 828.930698] env[61868]: DEBUG nova.compute.manager [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 828.930769] env[61868]: DEBUG nova.network.neutron [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 828.948958] env[61868]: DEBUG nova.network.neutron [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 829.036702] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "refresh_cache-a16a726b-9673-4de7-9188-19c608a7dc0b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.036702] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquired lock "refresh_cache-a16a726b-9673-4de7-9188-19c608a7dc0b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.036775] env[61868]: DEBUG nova.network.neutron [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 829.075177] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dda07a5d-b634-42f5-9075-a32c6a7a8886 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.083984] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6492de3e-390a-40bd-ad62-14a69a0c6c19 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.115056] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f42667d7-d848-4c33-b2c0-85ec7093b5b4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.122884] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-effd91d9-6f99-410a-8a03-16f45ab2119d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.141236] env[61868]: DEBUG nova.compute.provider_tree [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 829.148151] env[61868]: DEBUG oslo_vmware.api [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315357, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070924} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.148424] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 829.149191] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d8eeeb7-63ff-4a7b-af3c-01ecb71cfd98 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.169208] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Reconfiguring VM instance instance-00000038 to attach disk [datastore1] 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02/02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 829.170160] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-004c6428-bee5-45b8-a389-fed7434bb7e1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.190581] env[61868]: DEBUG oslo_vmware.api [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Waiting for the task: (returnval){ [ 829.190581] env[61868]: value = "task-1315358" [ 829.190581] env[61868]: _type = "Task" [ 829.190581] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.199079] env[61868]: DEBUG oslo_vmware.api [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315358, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.452572] env[61868]: DEBUG nova.network.neutron [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.580929] env[61868]: DEBUG nova.network.neutron [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 829.644429] env[61868]: DEBUG nova.scheduler.client.report [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 829.703252] env[61868]: DEBUG oslo_vmware.api [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315358, 'name': ReconfigVM_Task, 'duration_secs': 0.339418} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.703545] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Reconfigured VM instance instance-00000038 to attach disk [datastore1] 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02/02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 829.704169] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-634f378e-bbe9-4e0a-ae29-d4e5f3b01a3c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.710510] env[61868]: DEBUG oslo_vmware.api [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Waiting for the task: (returnval){ [ 829.710510] env[61868]: value = "task-1315359" [ 829.710510] env[61868]: _type = "Task" [ 829.710510] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.718664] env[61868]: DEBUG oslo_vmware.api [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315359, 'name': Rename_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.734475] env[61868]: DEBUG nova.network.neutron [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Updating instance_info_cache with network_info: [{"id": "548fe1c2-07a5-45a2-b084-cc95216cdcad", "address": "fa:16:3e:99:ec:d5", "network": {"id": "5b4c5906-b873-4369-855b-e0898ad84812", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-254512637-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "812e115cc1114d0e8536924c4f5556f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap548fe1c2-07", "ovs_interfaceid": "548fe1c2-07a5-45a2-b084-cc95216cdcad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.955521] env[61868]: INFO nova.compute.manager [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] [instance: 5c1683e1-0c87-4d1b-ae21-e07e04d848b1] Took 1.02 seconds to deallocate network for instance. [ 830.010411] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Acquiring lock "f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.010637] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Lock "f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.149405] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.384s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.149975] env[61868]: DEBUG nova.compute.manager [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 830.152824] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.772s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.154552] env[61868]: INFO nova.compute.claims [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 830.221492] env[61868]: DEBUG oslo_vmware.api [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315359, 'name': Rename_Task, 'duration_secs': 0.123716} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.221798] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 830.222050] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7f7aa968-1848-449d-8ac2-8a96700fa0b9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.229210] env[61868]: DEBUG oslo_vmware.api [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Waiting for the task: (returnval){ [ 830.229210] env[61868]: value = "task-1315360" [ 830.229210] env[61868]: _type = "Task" [ 830.229210] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.240150] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Releasing lock "refresh_cache-a16a726b-9673-4de7-9188-19c608a7dc0b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.240571] env[61868]: DEBUG nova.compute.manager [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Instance network_info: |[{"id": "548fe1c2-07a5-45a2-b084-cc95216cdcad", "address": "fa:16:3e:99:ec:d5", "network": {"id": "5b4c5906-b873-4369-855b-e0898ad84812", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-254512637-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "812e115cc1114d0e8536924c4f5556f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap548fe1c2-07", "ovs_interfaceid": "548fe1c2-07a5-45a2-b084-cc95216cdcad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 830.240903] env[61868]: DEBUG oslo_vmware.api [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315360, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.241306] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:99:ec:d5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a92a4ffe-7939-4697-bf98-5b22e2c7feda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '548fe1c2-07a5-45a2-b084-cc95216cdcad', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 830.248871] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Creating folder: Project (812e115cc1114d0e8536924c4f5556f8). Parent ref: group-v281478. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 830.249221] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-36044035-128a-4a2c-8ac2-a2393d8144e5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.258904] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Created folder: Project (812e115cc1114d0e8536924c4f5556f8) in parent group-v281478. [ 830.259108] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Creating folder: Instances. Parent ref: group-v281505. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 830.259459] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-089ff3fb-6b23-474b-890e-61f9e70a0e2d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.267820] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Created folder: Instances in parent group-v281505. [ 830.267997] env[61868]: DEBUG oslo.service.loopingcall [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 830.268210] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 830.268484] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bd8d29ba-393c-4c7c-a0b1-1f2dbf1af0cc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.287660] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 830.287660] env[61868]: value = "task-1315363" [ 830.287660] env[61868]: _type = "Task" [ 830.287660] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.295520] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315363, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.391136] env[61868]: DEBUG nova.compute.manager [req-998d2a86-90db-44f9-b6f7-0e039c042f39 req-86c28094-5e9c-448c-b7b9-ff5de4ee1e45 service nova] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Received event network-changed-548fe1c2-07a5-45a2-b084-cc95216cdcad {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 830.391328] env[61868]: DEBUG nova.compute.manager [req-998d2a86-90db-44f9-b6f7-0e039c042f39 req-86c28094-5e9c-448c-b7b9-ff5de4ee1e45 service nova] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Refreshing instance network info cache due to event network-changed-548fe1c2-07a5-45a2-b084-cc95216cdcad. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 830.391599] env[61868]: DEBUG oslo_concurrency.lockutils [req-998d2a86-90db-44f9-b6f7-0e039c042f39 req-86c28094-5e9c-448c-b7b9-ff5de4ee1e45 service nova] Acquiring lock "refresh_cache-a16a726b-9673-4de7-9188-19c608a7dc0b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.391744] env[61868]: DEBUG oslo_concurrency.lockutils [req-998d2a86-90db-44f9-b6f7-0e039c042f39 req-86c28094-5e9c-448c-b7b9-ff5de4ee1e45 service nova] Acquired lock "refresh_cache-a16a726b-9673-4de7-9188-19c608a7dc0b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.391901] env[61868]: DEBUG nova.network.neutron [req-998d2a86-90db-44f9-b6f7-0e039c042f39 req-86c28094-5e9c-448c-b7b9-ff5de4ee1e45 service nova] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Refreshing network info cache for port 548fe1c2-07a5-45a2-b084-cc95216cdcad {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 830.662565] env[61868]: DEBUG nova.compute.utils [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 830.666064] env[61868]: DEBUG nova.compute.manager [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 830.666198] env[61868]: DEBUG nova.network.neutron [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 830.718860] env[61868]: DEBUG nova.policy [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '03ead7f08e934c4e9a1b39fbe08ed6e3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2cb049f497b14ac39a0316132f5802fc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 830.739228] env[61868]: DEBUG oslo_vmware.api [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315360, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.797046] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315363, 'name': CreateVM_Task, 'duration_secs': 0.397252} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.797046] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 830.797685] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.798352] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.798418] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 830.798694] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f7cfefe-a688-4508-a2da-035235248071 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.803571] env[61868]: DEBUG oslo_vmware.api [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 830.803571] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e3da78-2cbc-2e6f-10e1-363affa6d80f" [ 830.803571] env[61868]: _type = "Task" [ 830.803571] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.811522] env[61868]: DEBUG oslo_vmware.api [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e3da78-2cbc-2e6f-10e1-363affa6d80f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.995024] env[61868]: INFO nova.scheduler.client.report [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] Deleted allocations for instance 5c1683e1-0c87-4d1b-ae21-e07e04d848b1 [ 831.050487] env[61868]: DEBUG nova.network.neutron [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Successfully created port: 6b7fa3c1-e436-4fed-92f3-d40988d5c79c {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 831.167018] env[61868]: DEBUG nova.compute.manager [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 831.234558] env[61868]: DEBUG nova.network.neutron [req-998d2a86-90db-44f9-b6f7-0e039c042f39 req-86c28094-5e9c-448c-b7b9-ff5de4ee1e45 service nova] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Updated VIF entry in instance network info cache for port 548fe1c2-07a5-45a2-b084-cc95216cdcad. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 831.234558] env[61868]: DEBUG nova.network.neutron [req-998d2a86-90db-44f9-b6f7-0e039c042f39 req-86c28094-5e9c-448c-b7b9-ff5de4ee1e45 service nova] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Updating instance_info_cache with network_info: [{"id": "548fe1c2-07a5-45a2-b084-cc95216cdcad", "address": "fa:16:3e:99:ec:d5", "network": {"id": "5b4c5906-b873-4369-855b-e0898ad84812", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-254512637-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "812e115cc1114d0e8536924c4f5556f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap548fe1c2-07", "ovs_interfaceid": "548fe1c2-07a5-45a2-b084-cc95216cdcad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.241022] env[61868]: DEBUG oslo_vmware.api [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315360, 'name': PowerOnVM_Task, 'duration_secs': 0.631456} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.241228] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 831.241456] env[61868]: INFO nova.compute.manager [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Took 5.80 seconds to spawn the instance on the hypervisor. [ 831.241664] env[61868]: DEBUG nova.compute.manager [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 831.245165] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70ccc999-1844-433e-bf18-d0e05beb84d2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.315449] env[61868]: DEBUG oslo_vmware.api [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e3da78-2cbc-2e6f-10e1-363affa6d80f, 'name': SearchDatastore_Task, 'duration_secs': 0.009695} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.318089] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.318328] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 831.318556] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.318700] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.318870] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 831.319303] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ddd154d5-bee3-4c82-bd0f-32caec90c99b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.332241] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 831.332408] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 831.333177] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60c1e66a-1379-4dfb-ad24-dfb1ea2ef7a9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.339257] env[61868]: DEBUG oslo_vmware.api [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 831.339257] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5215fa1f-eba5-3421-8150-7cdfae0e3927" [ 831.339257] env[61868]: _type = "Task" [ 831.339257] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.348723] env[61868]: DEBUG oslo_vmware.api [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5215fa1f-eba5-3421-8150-7cdfae0e3927, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.470966] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7212f09-111d-47e3-a94a-7573e78a4f66 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.478677] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bb4f3c6-59f9-4fa6-9af0-af91fde3c8fb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.507475] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b5ca76f6-61d2-4245-b9d8-06db238d79a1 tempest-ServerExternalEventsTest-311120656 tempest-ServerExternalEventsTest-311120656-project-member] Lock "5c1683e1-0c87-4d1b-ae21-e07e04d848b1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 142.719s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.508213] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cbc3e9a-d6d4-40ce-9e9e-b5f227a8d487 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.518170] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-551b7d79-8a52-4fcf-a9a6-ca8a6f5cdb10 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.530017] env[61868]: DEBUG nova.compute.provider_tree [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 831.737104] env[61868]: DEBUG oslo_concurrency.lockutils [req-998d2a86-90db-44f9-b6f7-0e039c042f39 req-86c28094-5e9c-448c-b7b9-ff5de4ee1e45 service nova] Releasing lock "refresh_cache-a16a726b-9673-4de7-9188-19c608a7dc0b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.764108] env[61868]: INFO nova.compute.manager [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Took 39.02 seconds to build instance. [ 831.851017] env[61868]: DEBUG oslo_vmware.api [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5215fa1f-eba5-3421-8150-7cdfae0e3927, 'name': SearchDatastore_Task, 'duration_secs': 0.009311} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.851017] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-278e279b-6df3-41e1-ba1e-04c70dd5e60e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.854795] env[61868]: DEBUG oslo_vmware.api [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 831.854795] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52564b1f-de5b-2818-72d0-d4d2bbb33116" [ 831.854795] env[61868]: _type = "Task" [ 831.854795] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.861903] env[61868]: DEBUG oslo_vmware.api [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52564b1f-de5b-2818-72d0-d4d2bbb33116, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.014254] env[61868]: DEBUG nova.compute.manager [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 832.033649] env[61868]: DEBUG nova.scheduler.client.report [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 832.181374] env[61868]: DEBUG nova.compute.manager [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 832.202757] env[61868]: DEBUG nova.virt.hardware [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 832.202757] env[61868]: DEBUG nova.virt.hardware [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 832.202936] env[61868]: DEBUG nova.virt.hardware [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 832.203102] env[61868]: DEBUG nova.virt.hardware [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 832.203228] env[61868]: DEBUG nova.virt.hardware [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 832.203375] env[61868]: DEBUG nova.virt.hardware [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 832.204030] env[61868]: DEBUG nova.virt.hardware [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 832.204030] env[61868]: DEBUG nova.virt.hardware [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 832.204030] env[61868]: DEBUG nova.virt.hardware [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 832.204162] env[61868]: DEBUG nova.virt.hardware [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 832.204221] env[61868]: DEBUG nova.virt.hardware [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 832.205071] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fc1180a-00d7-4899-9e26-3581ba03e697 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.214312] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55fd0c98-8511-4dbc-a763-3effaf3a7fa8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.265974] env[61868]: DEBUG oslo_concurrency.lockutils [None req-cd4d7a7a-f807-4563-9995-b67d2f045bb5 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Lock "02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 118.777s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.366232] env[61868]: DEBUG oslo_vmware.api [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52564b1f-de5b-2818-72d0-d4d2bbb33116, 'name': SearchDatastore_Task, 'duration_secs': 0.009471} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.366552] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.367235] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] a16a726b-9673-4de7-9188-19c608a7dc0b/a16a726b-9673-4de7-9188-19c608a7dc0b.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 832.367235] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a88cb03c-bf91-49d1-8b35-ee116e9c598c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.374540] env[61868]: DEBUG oslo_vmware.api [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 832.374540] env[61868]: value = "task-1315364" [ 832.374540] env[61868]: _type = "Task" [ 832.374540] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.382905] env[61868]: DEBUG oslo_vmware.api [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315364, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.538886] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.539205] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.386s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.539728] env[61868]: DEBUG nova.compute.manager [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 832.542305] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.514s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.543730] env[61868]: INFO nova.compute.claims [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 832.761746] env[61868]: DEBUG nova.compute.manager [req-1cb401fe-6f2e-43e7-a73f-b764a4c12afa req-e4702478-34e8-4ea7-828d-ebcd50a42235 service nova] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Received event network-vif-plugged-6b7fa3c1-e436-4fed-92f3-d40988d5c79c {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 832.761985] env[61868]: DEBUG oslo_concurrency.lockutils [req-1cb401fe-6f2e-43e7-a73f-b764a4c12afa req-e4702478-34e8-4ea7-828d-ebcd50a42235 service nova] Acquiring lock "0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.762159] env[61868]: DEBUG oslo_concurrency.lockutils [req-1cb401fe-6f2e-43e7-a73f-b764a4c12afa req-e4702478-34e8-4ea7-828d-ebcd50a42235 service nova] Lock "0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.762324] env[61868]: DEBUG oslo_concurrency.lockutils [req-1cb401fe-6f2e-43e7-a73f-b764a4c12afa req-e4702478-34e8-4ea7-828d-ebcd50a42235 service nova] Lock "0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.762484] env[61868]: DEBUG nova.compute.manager [req-1cb401fe-6f2e-43e7-a73f-b764a4c12afa req-e4702478-34e8-4ea7-828d-ebcd50a42235 service nova] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] No waiting events found dispatching network-vif-plugged-6b7fa3c1-e436-4fed-92f3-d40988d5c79c {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 832.762675] env[61868]: WARNING nova.compute.manager [req-1cb401fe-6f2e-43e7-a73f-b764a4c12afa req-e4702478-34e8-4ea7-828d-ebcd50a42235 service nova] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Received unexpected event network-vif-plugged-6b7fa3c1-e436-4fed-92f3-d40988d5c79c for instance with vm_state building and task_state spawning. [ 832.768972] env[61868]: DEBUG nova.compute.manager [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 832.801344] env[61868]: INFO nova.compute.manager [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Rebuilding instance [ 832.848753] env[61868]: DEBUG nova.compute.manager [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 832.849813] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54ab72ac-4b71-449b-bc4e-d1179851aa66 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.874910] env[61868]: DEBUG nova.network.neutron [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Successfully updated port: 6b7fa3c1-e436-4fed-92f3-d40988d5c79c {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 832.886841] env[61868]: DEBUG oslo_vmware.api [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315364, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.050472] env[61868]: DEBUG nova.compute.utils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 833.052117] env[61868]: DEBUG nova.compute.manager [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 833.052397] env[61868]: DEBUG nova.network.neutron [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 833.093308] env[61868]: DEBUG nova.policy [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '919b1476fe6d4aa7b4663165492ea564', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8cac586898424ba3a8c64d9156a68f86', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 833.297926] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.362820] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 833.366401] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-caebcdb4-fd9f-416f-8370-a16fc355478e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.374826] env[61868]: DEBUG oslo_vmware.api [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Waiting for the task: (returnval){ [ 833.374826] env[61868]: value = "task-1315365" [ 833.374826] env[61868]: _type = "Task" [ 833.374826] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.380855] env[61868]: DEBUG nova.network.neutron [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Successfully created port: ec4c8961-6ef8-4b0e-bb38-99c843e9fd4c {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 833.385196] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquiring lock "refresh_cache-0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.385351] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquired lock "refresh_cache-0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.385484] env[61868]: DEBUG nova.network.neutron [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 833.397304] env[61868]: DEBUG oslo_vmware.api [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315365, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.400509] env[61868]: DEBUG oslo_vmware.api [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315364, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.556281] env[61868]: DEBUG nova.compute.manager [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 833.814541] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cffd558b-0a76-42be-97cb-592577f6edb9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.822185] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38433978-97fd-424a-a5fe-e24629f49ffd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.853925] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32972462-af27-4210-852c-62a50979e023 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.863245] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bad5ef93-0594-4225-ad8e-a2291e07cfb4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.881375] env[61868]: DEBUG nova.compute.provider_tree [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 833.897727] env[61868]: DEBUG oslo_vmware.api [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315365, 'name': PowerOffVM_Task, 'duration_secs': 0.117383} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.901567] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 833.901803] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 833.902093] env[61868]: DEBUG oslo_vmware.api [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315364, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.903199] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f127b48b-bada-42d5-9420-dfc73d646ca7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.910667] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 833.910917] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e52fe621-16b4-4f6e-afb7-2bd9b58be601 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.934520] env[61868]: DEBUG nova.network.neutron [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 833.939864] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 833.939864] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 833.939864] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Deleting the datastore file [datastore1] 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 833.939864] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c727766f-d5bb-496c-976b-d1639b289529 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.946942] env[61868]: DEBUG oslo_vmware.api [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Waiting for the task: (returnval){ [ 833.946942] env[61868]: value = "task-1315367" [ 833.946942] env[61868]: _type = "Task" [ 833.946942] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.953696] env[61868]: DEBUG oslo_vmware.api [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315367, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.125617] env[61868]: DEBUG nova.network.neutron [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Updating instance_info_cache with network_info: [{"id": "6b7fa3c1-e436-4fed-92f3-d40988d5c79c", "address": "fa:16:3e:c3:b2:a6", "network": {"id": "369e3d75-3302-4816-b5f9-5da97f617152", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-855674976-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2cb049f497b14ac39a0316132f5802fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b7fa3c1-e4", "ovs_interfaceid": "6b7fa3c1-e436-4fed-92f3-d40988d5c79c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.387145] env[61868]: DEBUG nova.scheduler.client.report [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 834.401696] env[61868]: DEBUG oslo_vmware.api [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315364, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.456667] env[61868]: DEBUG oslo_vmware.api [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315367, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141456} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.456942] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 834.457150] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 834.457328] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 834.568180] env[61868]: DEBUG nova.compute.manager [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 834.593107] env[61868]: DEBUG nova.virt.hardware [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 834.593402] env[61868]: DEBUG nova.virt.hardware [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 834.593579] env[61868]: DEBUG nova.virt.hardware [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 834.593791] env[61868]: DEBUG nova.virt.hardware [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 834.593928] env[61868]: DEBUG nova.virt.hardware [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 834.594180] env[61868]: DEBUG nova.virt.hardware [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 834.594438] env[61868]: DEBUG nova.virt.hardware [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 834.594658] env[61868]: DEBUG nova.virt.hardware [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 834.594866] env[61868]: DEBUG nova.virt.hardware [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 834.595081] env[61868]: DEBUG nova.virt.hardware [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 834.595301] env[61868]: DEBUG nova.virt.hardware [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 834.596247] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9473db98-6386-4778-81b8-0407b28558cb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.604763] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d83992c8-cae3-4267-b991-db285840f2fd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.629383] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Releasing lock "refresh_cache-0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.629792] env[61868]: DEBUG nova.compute.manager [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Instance network_info: |[{"id": "6b7fa3c1-e436-4fed-92f3-d40988d5c79c", "address": "fa:16:3e:c3:b2:a6", "network": {"id": "369e3d75-3302-4816-b5f9-5da97f617152", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-855674976-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2cb049f497b14ac39a0316132f5802fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b7fa3c1-e4", "ovs_interfaceid": "6b7fa3c1-e436-4fed-92f3-d40988d5c79c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 834.630557] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c3:b2:a6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a8b99a46-3e7f-4ef1-9e45-58e6cd17f210', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6b7fa3c1-e436-4fed-92f3-d40988d5c79c', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 834.637815] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Creating folder: Project (2cb049f497b14ac39a0316132f5802fc). Parent ref: group-v281478. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 834.638122] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3f707d44-bdf0-4cef-a6fb-5a008e8649dd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.647674] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Created folder: Project (2cb049f497b14ac39a0316132f5802fc) in parent group-v281478. [ 834.647917] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Creating folder: Instances. Parent ref: group-v281508. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 834.648194] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-49097fb7-0825-43cb-aeb0-66b823b93b30 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.655898] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Created folder: Instances in parent group-v281508. [ 834.656183] env[61868]: DEBUG oslo.service.loopingcall [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 834.656403] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 834.656627] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ed9adbde-058f-40c7-b7bf-90a6a4eb6b8d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.674728] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 834.674728] env[61868]: value = "task-1315370" [ 834.674728] env[61868]: _type = "Task" [ 834.674728] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.682291] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315370, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.792545] env[61868]: DEBUG nova.compute.manager [req-29c34565-da85-4f69-b622-9dd0db5cd877 req-fe0eef39-fc75-4ef5-8e72-928b7ce0fddb service nova] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Received event network-changed-6b7fa3c1-e436-4fed-92f3-d40988d5c79c {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 834.792762] env[61868]: DEBUG nova.compute.manager [req-29c34565-da85-4f69-b622-9dd0db5cd877 req-fe0eef39-fc75-4ef5-8e72-928b7ce0fddb service nova] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Refreshing instance network info cache due to event network-changed-6b7fa3c1-e436-4fed-92f3-d40988d5c79c. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 834.793189] env[61868]: DEBUG oslo_concurrency.lockutils [req-29c34565-da85-4f69-b622-9dd0db5cd877 req-fe0eef39-fc75-4ef5-8e72-928b7ce0fddb service nova] Acquiring lock "refresh_cache-0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.793189] env[61868]: DEBUG oslo_concurrency.lockutils [req-29c34565-da85-4f69-b622-9dd0db5cd877 req-fe0eef39-fc75-4ef5-8e72-928b7ce0fddb service nova] Acquired lock "refresh_cache-0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.793314] env[61868]: DEBUG nova.network.neutron [req-29c34565-da85-4f69-b622-9dd0db5cd877 req-fe0eef39-fc75-4ef5-8e72-928b7ce0fddb service nova] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Refreshing network info cache for port 6b7fa3c1-e436-4fed-92f3-d40988d5c79c {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 834.896666] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.354s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.897379] env[61868]: DEBUG nova.compute.manager [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 834.904212] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.795s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.905739] env[61868]: INFO nova.compute.claims [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 834.908451] env[61868]: DEBUG oslo_vmware.api [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315364, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.111129} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.909066] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] a16a726b-9673-4de7-9188-19c608a7dc0b/a16a726b-9673-4de7-9188-19c608a7dc0b.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 834.910869] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 834.911179] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b8f39906-913d-4d90-a574-25bdf722b1c6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.925374] env[61868]: DEBUG oslo_vmware.api [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 834.925374] env[61868]: value = "task-1315371" [ 834.925374] env[61868]: _type = "Task" [ 834.925374] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.936340] env[61868]: DEBUG oslo_vmware.api [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315371, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.081449] env[61868]: DEBUG nova.network.neutron [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Successfully updated port: ec4c8961-6ef8-4b0e-bb38-99c843e9fd4c {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 835.183991] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315370, 'name': CreateVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.412607] env[61868]: DEBUG nova.compute.utils [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 835.413871] env[61868]: DEBUG nova.compute.manager [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 835.415979] env[61868]: DEBUG nova.network.neutron [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 7788f9f8-c08e-46f8-b204-070215223436] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 835.434350] env[61868]: DEBUG oslo_vmware.api [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315371, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.214635} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.434610] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 835.435382] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ff1a364-07d9-4df3-b2cc-740e29612dd6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.457437] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Reconfiguring VM instance instance-00000039 to attach disk [datastore1] a16a726b-9673-4de7-9188-19c608a7dc0b/a16a726b-9673-4de7-9188-19c608a7dc0b.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 835.460196] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-285a0330-96be-4049-90ef-ffc828daf3d0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.476933] env[61868]: DEBUG nova.policy [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '03ead7f08e934c4e9a1b39fbe08ed6e3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2cb049f497b14ac39a0316132f5802fc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 835.486503] env[61868]: DEBUG oslo_vmware.api [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 835.486503] env[61868]: value = "task-1315372" [ 835.486503] env[61868]: _type = "Task" [ 835.486503] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.494755] env[61868]: DEBUG oslo_vmware.api [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315372, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.504634] env[61868]: DEBUG nova.virt.hardware [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 835.504882] env[61868]: DEBUG nova.virt.hardware [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 835.505054] env[61868]: DEBUG nova.virt.hardware [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 835.505243] env[61868]: DEBUG nova.virt.hardware [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 835.505387] env[61868]: DEBUG nova.virt.hardware [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 835.505532] env[61868]: DEBUG nova.virt.hardware [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 835.505821] env[61868]: DEBUG nova.virt.hardware [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 835.505879] env[61868]: DEBUG nova.virt.hardware [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 835.506115] env[61868]: DEBUG nova.virt.hardware [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 835.506309] env[61868]: DEBUG nova.virt.hardware [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 835.506895] env[61868]: DEBUG nova.virt.hardware [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 835.507796] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8bb5966-eef7-4b6a-9502-9df08b5bc083 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.515831] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f223a91-2981-4ad5-b90c-2c1e59264eb4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.532434] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Instance VIF info [] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 835.538083] env[61868]: DEBUG oslo.service.loopingcall [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 835.538298] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 835.538513] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-25e1aefc-a5af-4f1d-b6de-b525031558dd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.556236] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 835.556236] env[61868]: value = "task-1315373" [ 835.556236] env[61868]: _type = "Task" [ 835.556236] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.563839] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315373, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.587237] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Acquiring lock "refresh_cache-3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.587237] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Acquired lock "refresh_cache-3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.587761] env[61868]: DEBUG nova.network.neutron [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 835.617874] env[61868]: DEBUG nova.network.neutron [req-29c34565-da85-4f69-b622-9dd0db5cd877 req-fe0eef39-fc75-4ef5-8e72-928b7ce0fddb service nova] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Updated VIF entry in instance network info cache for port 6b7fa3c1-e436-4fed-92f3-d40988d5c79c. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 835.619289] env[61868]: DEBUG nova.network.neutron [req-29c34565-da85-4f69-b622-9dd0db5cd877 req-fe0eef39-fc75-4ef5-8e72-928b7ce0fddb service nova] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Updating instance_info_cache with network_info: [{"id": "6b7fa3c1-e436-4fed-92f3-d40988d5c79c", "address": "fa:16:3e:c3:b2:a6", "network": {"id": "369e3d75-3302-4816-b5f9-5da97f617152", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-855674976-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2cb049f497b14ac39a0316132f5802fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b7fa3c1-e4", "ovs_interfaceid": "6b7fa3c1-e436-4fed-92f3-d40988d5c79c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.687023] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315370, 'name': CreateVM_Task, 'duration_secs': 0.590986} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.687023] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 835.687023] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.687023] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.687023] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 835.687023] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74d976d0-fea9-422e-93a8-8fcc53a8a99d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.691549] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Waiting for the task: (returnval){ [ 835.691549] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]527ff179-5f77-d117-bc32-7f9cf826122c" [ 835.691549] env[61868]: _type = "Task" [ 835.691549] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.700290] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]527ff179-5f77-d117-bc32-7f9cf826122c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.797536] env[61868]: DEBUG nova.network.neutron [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Successfully created port: 29d0d831-eb07-403f-bde9-16844b987447 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 835.918913] env[61868]: DEBUG nova.compute.manager [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 835.995254] env[61868]: DEBUG oslo_vmware.api [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315372, 'name': ReconfigVM_Task, 'duration_secs': 0.260006} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.997774] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Reconfigured VM instance instance-00000039 to attach disk [datastore1] a16a726b-9673-4de7-9188-19c608a7dc0b/a16a726b-9673-4de7-9188-19c608a7dc0b.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 835.998606] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6f9b6493-4b0d-4cfb-b376-0f5df0913924 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.004838] env[61868]: DEBUG oslo_vmware.api [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 836.004838] env[61868]: value = "task-1315374" [ 836.004838] env[61868]: _type = "Task" [ 836.004838] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.014870] env[61868]: DEBUG oslo_vmware.api [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315374, 'name': Rename_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.075731] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315373, 'name': CreateVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.121526] env[61868]: DEBUG oslo_concurrency.lockutils [req-29c34565-da85-4f69-b622-9dd0db5cd877 req-fe0eef39-fc75-4ef5-8e72-928b7ce0fddb service nova] Releasing lock "refresh_cache-0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.125940] env[61868]: DEBUG nova.network.neutron [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 836.209236] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]527ff179-5f77-d117-bc32-7f9cf826122c, 'name': SearchDatastore_Task, 'duration_secs': 0.010855} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.209537] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.209762] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 836.209983] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.210133] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.210301] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 836.210564] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ff42cf36-5e18-4e4b-a509-4bb8582531c7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.219907] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 836.220101] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 836.220829] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d93dbe81-2f9d-458b-8d30-983602f8671c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.226497] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Waiting for the task: (returnval){ [ 836.226497] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]528467eb-0733-d41e-da1e-e0e1392f2c3c" [ 836.226497] env[61868]: _type = "Task" [ 836.226497] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.237678] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]528467eb-0733-d41e-da1e-e0e1392f2c3c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.289229] env[61868]: DEBUG nova.network.neutron [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Updating instance_info_cache with network_info: [{"id": "ec4c8961-6ef8-4b0e-bb38-99c843e9fd4c", "address": "fa:16:3e:4d:a1:b6", "network": {"id": "57a3a2e2-7168-4e6c-b8e7-d4904576ffea", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1599066480-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8cac586898424ba3a8c64d9156a68f86", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1430a695-49fb-4905-bc38-db9b869a1a9d", "external-id": "nsx-vlan-transportzone-297", "segmentation_id": 297, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec4c8961-6e", "ovs_interfaceid": "ec4c8961-6ef8-4b0e-bb38-99c843e9fd4c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.323859] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07fce99b-854d-46bf-ad95-204f77800353 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.333470] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fae4e13-cd08-4a1d-a5b4-d93113fe50c7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.363047] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-852d3714-c31d-4eaa-8ea3-280bd453ac44 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.369886] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67cfe79e-3a39-4122-a1cf-54ecf18ae683 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.384290] env[61868]: DEBUG nova.compute.provider_tree [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 836.514856] env[61868]: DEBUG oslo_vmware.api [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315374, 'name': Rename_Task, 'duration_secs': 0.128539} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.515199] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 836.515440] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-89ef50f6-f6df-46cf-9ea0-c593d51cc4cf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.522431] env[61868]: DEBUG oslo_vmware.api [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 836.522431] env[61868]: value = "task-1315375" [ 836.522431] env[61868]: _type = "Task" [ 836.522431] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.529795] env[61868]: DEBUG oslo_vmware.api [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315375, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.567123] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315373, 'name': CreateVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.738198] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]528467eb-0733-d41e-da1e-e0e1392f2c3c, 'name': SearchDatastore_Task, 'duration_secs': 0.008581} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.738980] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-689e9cff-4665-495a-bdab-e762fe63c3c8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.744611] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Waiting for the task: (returnval){ [ 836.744611] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52482495-653d-f5d4-d996-432a907c3b6a" [ 836.744611] env[61868]: _type = "Task" [ 836.744611] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.752912] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52482495-653d-f5d4-d996-432a907c3b6a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.795372] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Releasing lock "refresh_cache-3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.795372] env[61868]: DEBUG nova.compute.manager [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Instance network_info: |[{"id": "ec4c8961-6ef8-4b0e-bb38-99c843e9fd4c", "address": "fa:16:3e:4d:a1:b6", "network": {"id": "57a3a2e2-7168-4e6c-b8e7-d4904576ffea", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1599066480-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8cac586898424ba3a8c64d9156a68f86", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1430a695-49fb-4905-bc38-db9b869a1a9d", "external-id": "nsx-vlan-transportzone-297", "segmentation_id": 297, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec4c8961-6e", "ovs_interfaceid": "ec4c8961-6ef8-4b0e-bb38-99c843e9fd4c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 836.795582] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4d:a1:b6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1430a695-49fb-4905-bc38-db9b869a1a9d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ec4c8961-6ef8-4b0e-bb38-99c843e9fd4c', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 836.803472] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Creating folder: Project (8cac586898424ba3a8c64d9156a68f86). Parent ref: group-v281478. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 836.803472] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c379a266-4423-4ad9-80ec-bb16b65a2715 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.813322] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Created folder: Project (8cac586898424ba3a8c64d9156a68f86) in parent group-v281478. [ 836.814024] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Creating folder: Instances. Parent ref: group-v281512. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 836.814024] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a973accd-cfd6-4f44-ad99-95ab03706e39 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.821029] env[61868]: DEBUG nova.compute.manager [req-d2c7bf0f-b17f-4910-a372-21c283cca565 req-d718de0d-8be7-4650-865d-814a3e35d8b3 service nova] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Received event network-vif-plugged-ec4c8961-6ef8-4b0e-bb38-99c843e9fd4c {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 836.821257] env[61868]: DEBUG oslo_concurrency.lockutils [req-d2c7bf0f-b17f-4910-a372-21c283cca565 req-d718de0d-8be7-4650-865d-814a3e35d8b3 service nova] Acquiring lock "3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.821643] env[61868]: DEBUG oslo_concurrency.lockutils [req-d2c7bf0f-b17f-4910-a372-21c283cca565 req-d718de0d-8be7-4650-865d-814a3e35d8b3 service nova] Lock "3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.821763] env[61868]: DEBUG oslo_concurrency.lockutils [req-d2c7bf0f-b17f-4910-a372-21c283cca565 req-d718de0d-8be7-4650-865d-814a3e35d8b3 service nova] Lock "3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.822037] env[61868]: DEBUG nova.compute.manager [req-d2c7bf0f-b17f-4910-a372-21c283cca565 req-d718de0d-8be7-4650-865d-814a3e35d8b3 service nova] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] No waiting events found dispatching network-vif-plugged-ec4c8961-6ef8-4b0e-bb38-99c843e9fd4c {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 836.822316] env[61868]: WARNING nova.compute.manager [req-d2c7bf0f-b17f-4910-a372-21c283cca565 req-d718de0d-8be7-4650-865d-814a3e35d8b3 service nova] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Received unexpected event network-vif-plugged-ec4c8961-6ef8-4b0e-bb38-99c843e9fd4c for instance with vm_state building and task_state spawning. [ 836.822316] env[61868]: DEBUG nova.compute.manager [req-d2c7bf0f-b17f-4910-a372-21c283cca565 req-d718de0d-8be7-4650-865d-814a3e35d8b3 service nova] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Received event network-changed-ec4c8961-6ef8-4b0e-bb38-99c843e9fd4c {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 836.822416] env[61868]: DEBUG nova.compute.manager [req-d2c7bf0f-b17f-4910-a372-21c283cca565 req-d718de0d-8be7-4650-865d-814a3e35d8b3 service nova] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Refreshing instance network info cache due to event network-changed-ec4c8961-6ef8-4b0e-bb38-99c843e9fd4c. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 836.822695] env[61868]: DEBUG oslo_concurrency.lockutils [req-d2c7bf0f-b17f-4910-a372-21c283cca565 req-d718de0d-8be7-4650-865d-814a3e35d8b3 service nova] Acquiring lock "refresh_cache-3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.822747] env[61868]: DEBUG oslo_concurrency.lockutils [req-d2c7bf0f-b17f-4910-a372-21c283cca565 req-d718de0d-8be7-4650-865d-814a3e35d8b3 service nova] Acquired lock "refresh_cache-3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.822891] env[61868]: DEBUG nova.network.neutron [req-d2c7bf0f-b17f-4910-a372-21c283cca565 req-d718de0d-8be7-4650-865d-814a3e35d8b3 service nova] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Refreshing network info cache for port ec4c8961-6ef8-4b0e-bb38-99c843e9fd4c {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 836.826401] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Created folder: Instances in parent group-v281512. [ 836.827782] env[61868]: DEBUG oslo.service.loopingcall [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 836.827782] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 836.827782] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bce536bb-4db4-420e-a37c-3b2d422e11e1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.853321] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 836.853321] env[61868]: value = "task-1315378" [ 836.853321] env[61868]: _type = "Task" [ 836.853321] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.861823] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315378, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.888212] env[61868]: DEBUG nova.scheduler.client.report [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 836.936062] env[61868]: DEBUG nova.compute.manager [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 836.958302] env[61868]: DEBUG nova.virt.hardware [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 836.958860] env[61868]: DEBUG nova.virt.hardware [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 836.958860] env[61868]: DEBUG nova.virt.hardware [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 836.958971] env[61868]: DEBUG nova.virt.hardware [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 836.959070] env[61868]: DEBUG nova.virt.hardware [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 836.959219] env[61868]: DEBUG nova.virt.hardware [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 836.959433] env[61868]: DEBUG nova.virt.hardware [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 836.959626] env[61868]: DEBUG nova.virt.hardware [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 836.959811] env[61868]: DEBUG nova.virt.hardware [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 836.959986] env[61868]: DEBUG nova.virt.hardware [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 836.960193] env[61868]: DEBUG nova.virt.hardware [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 836.961071] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b141140-e64e-460f-83e4-574329659ace {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.968812] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b303af5-b81b-47e3-9419-5821d776c5a0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.032787] env[61868]: DEBUG oslo_vmware.api [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315375, 'name': PowerOnVM_Task, 'duration_secs': 0.418032} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.033058] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 837.033281] env[61868]: INFO nova.compute.manager [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Took 9.14 seconds to spawn the instance on the hypervisor. [ 837.033486] env[61868]: DEBUG nova.compute.manager [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 837.034666] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-205257d2-e290-42df-998c-7c1ad6c9d72e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.066413] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315373, 'name': CreateVM_Task, 'duration_secs': 1.32757} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.066598] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 837.067081] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.067254] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.067576] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 837.067828] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c28525b0-f355-4ab7-abb4-87e52919ff83 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.072530] env[61868]: DEBUG oslo_vmware.api [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Waiting for the task: (returnval){ [ 837.072530] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]523ea802-d5e0-b677-7eeb-4ac30061ef1e" [ 837.072530] env[61868]: _type = "Task" [ 837.072530] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.083869] env[61868]: DEBUG oslo_vmware.api [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]523ea802-d5e0-b677-7eeb-4ac30061ef1e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.257168] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52482495-653d-f5d4-d996-432a907c3b6a, 'name': SearchDatastore_Task, 'duration_secs': 0.009072} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.257486] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.257674] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc/0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 837.257925] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9c68d7cd-f28a-4cbe-a3f1-96eaa004da27 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.265456] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Waiting for the task: (returnval){ [ 837.265456] env[61868]: value = "task-1315379" [ 837.265456] env[61868]: _type = "Task" [ 837.265456] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.277774] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': task-1315379, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.364617] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315378, 'name': CreateVM_Task, 'duration_secs': 0.328127} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.364785] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 837.368037] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.380913] env[61868]: DEBUG nova.network.neutron [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Successfully updated port: 29d0d831-eb07-403f-bde9-16844b987447 {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 837.392944] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.489s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.393702] env[61868]: DEBUG nova.compute.manager [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 837.396775] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.173s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.399503] env[61868]: INFO nova.compute.claims [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 837.409031] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "306f2dc4-933c-4948-8330-eda93931c4e0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.410078] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "306f2dc4-933c-4948-8330-eda93931c4e0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.555112] env[61868]: INFO nova.compute.manager [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Took 42.88 seconds to build instance. [ 837.574169] env[61868]: DEBUG nova.network.neutron [req-d2c7bf0f-b17f-4910-a372-21c283cca565 req-d718de0d-8be7-4650-865d-814a3e35d8b3 service nova] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Updated VIF entry in instance network info cache for port ec4c8961-6ef8-4b0e-bb38-99c843e9fd4c. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 837.574169] env[61868]: DEBUG nova.network.neutron [req-d2c7bf0f-b17f-4910-a372-21c283cca565 req-d718de0d-8be7-4650-865d-814a3e35d8b3 service nova] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Updating instance_info_cache with network_info: [{"id": "ec4c8961-6ef8-4b0e-bb38-99c843e9fd4c", "address": "fa:16:3e:4d:a1:b6", "network": {"id": "57a3a2e2-7168-4e6c-b8e7-d4904576ffea", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1599066480-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8cac586898424ba3a8c64d9156a68f86", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1430a695-49fb-4905-bc38-db9b869a1a9d", "external-id": "nsx-vlan-transportzone-297", "segmentation_id": 297, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec4c8961-6e", "ovs_interfaceid": "ec4c8961-6ef8-4b0e-bb38-99c843e9fd4c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.589223] env[61868]: DEBUG oslo_vmware.api [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]523ea802-d5e0-b677-7eeb-4ac30061ef1e, 'name': SearchDatastore_Task, 'duration_secs': 0.012756} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.589223] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.589223] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 837.589223] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.589563] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.589563] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 837.589563] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.589563] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 837.589563] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0a01f7e2-b6ad-4e5a-ab7d-c090683c434e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.591770] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2cc7b245-7f50-4550-b926-ebf2b67964e4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.598468] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Waiting for the task: (returnval){ [ 837.598468] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e50d16-fe64-58f1-f8b2-f34021cdc3a0" [ 837.598468] env[61868]: _type = "Task" [ 837.598468] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.607362] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 837.607582] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 837.610466] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38fb32d7-4a1f-4f45-b12a-597f71a43d99 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.618996] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e50d16-fe64-58f1-f8b2-f34021cdc3a0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.622566] env[61868]: DEBUG oslo_vmware.api [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Waiting for the task: (returnval){ [ 837.622566] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52021a80-da6e-779d-f8eb-4639f4eeffdc" [ 837.622566] env[61868]: _type = "Task" [ 837.622566] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.632152] env[61868]: DEBUG oslo_vmware.api [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52021a80-da6e-779d-f8eb-4639f4eeffdc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.776270] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': task-1315379, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.502792} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.776595] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc/0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 837.778197] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 837.778197] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-56ab242a-534b-4d8f-9662-af37869751de {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.783809] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Waiting for the task: (returnval){ [ 837.783809] env[61868]: value = "task-1315380" [ 837.783809] env[61868]: _type = "Task" [ 837.783809] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.793113] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': task-1315380, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.884975] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquiring lock "refresh_cache-7788f9f8-c08e-46f8-b204-070215223436" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.885159] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquired lock "refresh_cache-7788f9f8-c08e-46f8-b204-070215223436" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.885316] env[61868]: DEBUG nova.network.neutron [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 837.905151] env[61868]: DEBUG nova.compute.utils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 837.906720] env[61868]: DEBUG nova.compute.manager [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 837.907482] env[61868]: DEBUG nova.network.neutron [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 837.960491] env[61868]: DEBUG nova.policy [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '919b1476fe6d4aa7b4663165492ea564', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8cac586898424ba3a8c64d9156a68f86', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 838.057989] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5ea5531d-acc7-4330-8f4d-27083846ce72 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "a16a726b-9673-4de7-9188-19c608a7dc0b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 124.565s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.079019] env[61868]: DEBUG oslo_concurrency.lockutils [req-d2c7bf0f-b17f-4910-a372-21c283cca565 req-d718de0d-8be7-4650-865d-814a3e35d8b3 service nova] Releasing lock "refresh_cache-3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.110757] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e50d16-fe64-58f1-f8b2-f34021cdc3a0, 'name': SearchDatastore_Task, 'duration_secs': 0.025794} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.111072] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.111295] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 838.111500] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.134063] env[61868]: DEBUG oslo_vmware.api [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52021a80-da6e-779d-f8eb-4639f4eeffdc, 'name': SearchDatastore_Task, 'duration_secs': 0.019854} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.134898] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06494abb-69cd-42dc-a5b1-a23ab063ec77 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.140331] env[61868]: DEBUG oslo_vmware.api [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Waiting for the task: (returnval){ [ 838.140331] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]520dc111-29e3-0ae7-5843-58b0aae0dad7" [ 838.140331] env[61868]: _type = "Task" [ 838.140331] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.149521] env[61868]: DEBUG oslo_vmware.api [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]520dc111-29e3-0ae7-5843-58b0aae0dad7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.244154] env[61868]: DEBUG nova.network.neutron [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Successfully created port: de6238cd-1354-4052-9aff-0d28691de0c8 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 838.294221] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': task-1315380, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.05961} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.294498] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 838.295338] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1203e7a-9cff-485b-8974-ff93d6c5a37a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.318254] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc/0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 838.319020] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7d170e2b-9332-448e-8f2f-4626cb9e2b54 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.339178] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Waiting for the task: (returnval){ [ 838.339178] env[61868]: value = "task-1315381" [ 838.339178] env[61868]: _type = "Task" [ 838.339178] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.351544] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': task-1315381, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.410856] env[61868]: DEBUG nova.compute.manager [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 838.437876] env[61868]: DEBUG nova.network.neutron [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 838.560460] env[61868]: DEBUG nova.compute.manager [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 838.578063] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "a16a726b-9673-4de7-9188-19c608a7dc0b" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.579020] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "a16a726b-9673-4de7-9188-19c608a7dc0b" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.579020] env[61868]: INFO nova.compute.manager [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Shelving [ 838.653597] env[61868]: DEBUG oslo_vmware.api [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]520dc111-29e3-0ae7-5843-58b0aae0dad7, 'name': SearchDatastore_Task, 'duration_secs': 0.010175} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.653692] env[61868]: DEBUG nova.network.neutron [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Updating instance_info_cache with network_info: [{"id": "29d0d831-eb07-403f-bde9-16844b987447", "address": "fa:16:3e:ca:31:df", "network": {"id": "369e3d75-3302-4816-b5f9-5da97f617152", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-855674976-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2cb049f497b14ac39a0316132f5802fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap29d0d831-eb", "ovs_interfaceid": "29d0d831-eb07-403f-bde9-16844b987447", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.657593] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.657911] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02/02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 838.658695] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.658833] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 838.659125] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-310efa29-5806-4312-b830-ce821661ac99 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.661087] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-53d64a0d-3cfa-4213-aff9-fdd58dab27b2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.672237] env[61868]: DEBUG oslo_vmware.api [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Waiting for the task: (returnval){ [ 838.672237] env[61868]: value = "task-1315382" [ 838.672237] env[61868]: _type = "Task" [ 838.672237] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.673022] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 838.673022] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 838.674768] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d000aac7-6ee1-4144-8f09-66d6c302ffae {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.696016] env[61868]: DEBUG oslo_vmware.api [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315382, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.696016] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Waiting for the task: (returnval){ [ 838.696016] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52995067-a863-43c5-fcb8-1d56502befc2" [ 838.696016] env[61868]: _type = "Task" [ 838.696016] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.705658] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52995067-a863-43c5-fcb8-1d56502befc2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.796343] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-695eb000-0100-4064-b2cf-2c71a34822a7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.804363] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb86969b-443a-4474-a69b-05826a106450 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.835917] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9088ab61-327e-4241-99da-5168b6e25fd3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.850892] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a1fe4d1-37fd-4159-8653-b4ce5b905086 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.855073] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': task-1315381, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.866110] env[61868]: DEBUG nova.compute.provider_tree [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 838.890569] env[61868]: DEBUG nova.compute.manager [req-23fdc9d0-fa5c-4bfc-92db-f8d1f9d0d5a9 req-dcd0dcc2-f2c0-480f-9613-da97419bfb02 service nova] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Received event network-vif-plugged-29d0d831-eb07-403f-bde9-16844b987447 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 838.890980] env[61868]: DEBUG oslo_concurrency.lockutils [req-23fdc9d0-fa5c-4bfc-92db-f8d1f9d0d5a9 req-dcd0dcc2-f2c0-480f-9613-da97419bfb02 service nova] Acquiring lock "7788f9f8-c08e-46f8-b204-070215223436-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.891177] env[61868]: DEBUG oslo_concurrency.lockutils [req-23fdc9d0-fa5c-4bfc-92db-f8d1f9d0d5a9 req-dcd0dcc2-f2c0-480f-9613-da97419bfb02 service nova] Lock "7788f9f8-c08e-46f8-b204-070215223436-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.891421] env[61868]: DEBUG oslo_concurrency.lockutils [req-23fdc9d0-fa5c-4bfc-92db-f8d1f9d0d5a9 req-dcd0dcc2-f2c0-480f-9613-da97419bfb02 service nova] Lock "7788f9f8-c08e-46f8-b204-070215223436-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.891722] env[61868]: DEBUG nova.compute.manager [req-23fdc9d0-fa5c-4bfc-92db-f8d1f9d0d5a9 req-dcd0dcc2-f2c0-480f-9613-da97419bfb02 service nova] [instance: 7788f9f8-c08e-46f8-b204-070215223436] No waiting events found dispatching network-vif-plugged-29d0d831-eb07-403f-bde9-16844b987447 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 838.891924] env[61868]: WARNING nova.compute.manager [req-23fdc9d0-fa5c-4bfc-92db-f8d1f9d0d5a9 req-dcd0dcc2-f2c0-480f-9613-da97419bfb02 service nova] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Received unexpected event network-vif-plugged-29d0d831-eb07-403f-bde9-16844b987447 for instance with vm_state building and task_state spawning. [ 838.892105] env[61868]: DEBUG nova.compute.manager [req-23fdc9d0-fa5c-4bfc-92db-f8d1f9d0d5a9 req-dcd0dcc2-f2c0-480f-9613-da97419bfb02 service nova] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Received event network-changed-29d0d831-eb07-403f-bde9-16844b987447 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 838.892615] env[61868]: DEBUG nova.compute.manager [req-23fdc9d0-fa5c-4bfc-92db-f8d1f9d0d5a9 req-dcd0dcc2-f2c0-480f-9613-da97419bfb02 service nova] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Refreshing instance network info cache due to event network-changed-29d0d831-eb07-403f-bde9-16844b987447. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 838.892615] env[61868]: DEBUG oslo_concurrency.lockutils [req-23fdc9d0-fa5c-4bfc-92db-f8d1f9d0d5a9 req-dcd0dcc2-f2c0-480f-9613-da97419bfb02 service nova] Acquiring lock "refresh_cache-7788f9f8-c08e-46f8-b204-070215223436" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.086961] env[61868]: DEBUG oslo_concurrency.lockutils [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.087329] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 839.087595] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-47f48396-b09d-4fdc-abc5-ce9665260178 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.096083] env[61868]: DEBUG oslo_vmware.api [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 839.096083] env[61868]: value = "task-1315383" [ 839.096083] env[61868]: _type = "Task" [ 839.096083] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.106323] env[61868]: DEBUG oslo_vmware.api [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315383, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.159263] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Releasing lock "refresh_cache-7788f9f8-c08e-46f8-b204-070215223436" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.159659] env[61868]: DEBUG nova.compute.manager [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Instance network_info: |[{"id": "29d0d831-eb07-403f-bde9-16844b987447", "address": "fa:16:3e:ca:31:df", "network": {"id": "369e3d75-3302-4816-b5f9-5da97f617152", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-855674976-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2cb049f497b14ac39a0316132f5802fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap29d0d831-eb", "ovs_interfaceid": "29d0d831-eb07-403f-bde9-16844b987447", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 839.159999] env[61868]: DEBUG oslo_concurrency.lockutils [req-23fdc9d0-fa5c-4bfc-92db-f8d1f9d0d5a9 req-dcd0dcc2-f2c0-480f-9613-da97419bfb02 service nova] Acquired lock "refresh_cache-7788f9f8-c08e-46f8-b204-070215223436" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.160234] env[61868]: DEBUG nova.network.neutron [req-23fdc9d0-fa5c-4bfc-92db-f8d1f9d0d5a9 req-dcd0dcc2-f2c0-480f-9613-da97419bfb02 service nova] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Refreshing network info cache for port 29d0d831-eb07-403f-bde9-16844b987447 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 839.161555] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ca:31:df', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a8b99a46-3e7f-4ef1-9e45-58e6cd17f210', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '29d0d831-eb07-403f-bde9-16844b987447', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 839.169528] env[61868]: DEBUG oslo.service.loopingcall [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 839.169907] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 839.170151] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0f0bca02-c10f-4adb-97bd-2e08e6e07649 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.194671] env[61868]: DEBUG oslo_vmware.api [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315382, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.47147} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.196163] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02/02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 839.196382] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 839.196716] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 839.196716] env[61868]: value = "task-1315384" [ 839.196716] env[61868]: _type = "Task" [ 839.196716] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.199405] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-da242011-4008-4b27-92ce-e427bd17527d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.210140] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52995067-a863-43c5-fcb8-1d56502befc2, 'name': SearchDatastore_Task, 'duration_secs': 0.016866} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.213547] env[61868]: DEBUG oslo_vmware.api [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Waiting for the task: (returnval){ [ 839.213547] env[61868]: value = "task-1315385" [ 839.213547] env[61868]: _type = "Task" [ 839.213547] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.214160] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-174ff220-e752-424c-8d04-bf935ff02d78 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.223177] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315384, 'name': CreateVM_Task} progress is 10%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.228678] env[61868]: DEBUG oslo_vmware.api [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315385, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.229048] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Waiting for the task: (returnval){ [ 839.229048] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52d48cc0-32df-17cb-7456-5561ba9df215" [ 839.229048] env[61868]: _type = "Task" [ 839.229048] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.236383] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52d48cc0-32df-17cb-7456-5561ba9df215, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.353593] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': task-1315381, 'name': ReconfigVM_Task, 'duration_secs': 0.766151} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.354014] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Reconfigured VM instance instance-0000003a to attach disk [datastore1] 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc/0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 839.354742] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bff79ce9-8470-401c-a1e5-791203b5c2c1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.361658] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Waiting for the task: (returnval){ [ 839.361658] env[61868]: value = "task-1315386" [ 839.361658] env[61868]: _type = "Task" [ 839.361658] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.372708] env[61868]: DEBUG nova.scheduler.client.report [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 839.375785] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': task-1315386, 'name': Rename_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.423094] env[61868]: DEBUG nova.compute.manager [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 839.456127] env[61868]: DEBUG nova.virt.hardware [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 839.456127] env[61868]: DEBUG nova.virt.hardware [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 839.456127] env[61868]: DEBUG nova.virt.hardware [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 839.456343] env[61868]: DEBUG nova.virt.hardware [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 839.456343] env[61868]: DEBUG nova.virt.hardware [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 839.456343] env[61868]: DEBUG nova.virt.hardware [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 839.456812] env[61868]: DEBUG nova.virt.hardware [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 839.457215] env[61868]: DEBUG nova.virt.hardware [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 839.457539] env[61868]: DEBUG nova.virt.hardware [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 839.457923] env[61868]: DEBUG nova.virt.hardware [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 839.460020] env[61868]: DEBUG nova.virt.hardware [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 839.460020] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69125aad-b46c-4474-b4dc-e1e183996c12 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.470552] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78a7f0d4-93fb-4091-9691-91affdde97ff {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.610667] env[61868]: DEBUG oslo_vmware.api [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315383, 'name': PowerOffVM_Task, 'duration_secs': 0.298389} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.610667] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 839.611874] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ead6e65-6194-4bea-b0bd-e76f9d80cc74 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.635524] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e877c41a-95bb-4cbf-ae32-7bbbb9134db3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.715949] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315384, 'name': CreateVM_Task, 'duration_secs': 0.343131} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.715949] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 839.715949] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.715949] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.715949] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 839.715949] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eabde926-6454-42f8-9cef-ee9a020d35c0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.730792] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Waiting for the task: (returnval){ [ 839.730792] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]522ff9b8-1f12-48df-591a-9f92fd5a56fa" [ 839.730792] env[61868]: _type = "Task" [ 839.730792] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.737198] env[61868]: DEBUG oslo_vmware.api [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315385, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072083} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.740636] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 839.741703] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20564da0-d273-463e-962c-82166c014c1f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.749512] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52d48cc0-32df-17cb-7456-5561ba9df215, 'name': SearchDatastore_Task, 'duration_secs': 0.009653} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.750032] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.750291] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1/3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 839.753610] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4cf6929d-29bd-4a07-a1dc-47a10900daa3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.759019] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]522ff9b8-1f12-48df-591a-9f92fd5a56fa, 'name': SearchDatastore_Task, 'duration_secs': 0.01016} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.763691] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.763962] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 839.764223] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.764373] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.764549] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 839.774183] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Reconfiguring VM instance instance-00000038 to attach disk [datastore2] 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02/02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 839.774832] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5b989834-71ef-4408-bec0-e5b3578473f1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.778288] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3dacb148-d1a5-4ad5-b230-96fbb4b3fc1f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.798021] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Waiting for the task: (returnval){ [ 839.798021] env[61868]: value = "task-1315387" [ 839.798021] env[61868]: _type = "Task" [ 839.798021] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.802894] env[61868]: DEBUG oslo_vmware.api [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Waiting for the task: (returnval){ [ 839.802894] env[61868]: value = "task-1315388" [ 839.802894] env[61868]: _type = "Task" [ 839.802894] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.806592] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 839.806752] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 839.807461] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315387, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.810573] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1facba6-58f7-4ab1-853b-3a2c7d127a24 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.815573] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Waiting for the task: (returnval){ [ 839.815573] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]528671b8-5a8d-9347-11c8-61f03aabf847" [ 839.815573] env[61868]: _type = "Task" [ 839.815573] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.820491] env[61868]: DEBUG oslo_vmware.api [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315388, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.831775] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]528671b8-5a8d-9347-11c8-61f03aabf847, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.872280] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': task-1315386, 'name': Rename_Task, 'duration_secs': 0.249036} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.872280] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 839.872280] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7585c4d5-9f46-4a5c-a51b-e01dc7a8dd60 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.877413] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.481s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.877954] env[61868]: DEBUG nova.compute.manager [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 839.884030] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Waiting for the task: (returnval){ [ 839.884030] env[61868]: value = "task-1315389" [ 839.884030] env[61868]: _type = "Task" [ 839.884030] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.884030] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.063s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.884030] env[61868]: INFO nova.compute.claims [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 839.895947] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': task-1315389, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.040432] env[61868]: DEBUG nova.network.neutron [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Successfully updated port: de6238cd-1354-4052-9aff-0d28691de0c8 {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 840.053888] env[61868]: DEBUG nova.network.neutron [req-23fdc9d0-fa5c-4bfc-92db-f8d1f9d0d5a9 req-dcd0dcc2-f2c0-480f-9613-da97419bfb02 service nova] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Updated VIF entry in instance network info cache for port 29d0d831-eb07-403f-bde9-16844b987447. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 840.054350] env[61868]: DEBUG nova.network.neutron [req-23fdc9d0-fa5c-4bfc-92db-f8d1f9d0d5a9 req-dcd0dcc2-f2c0-480f-9613-da97419bfb02 service nova] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Updating instance_info_cache with network_info: [{"id": "29d0d831-eb07-403f-bde9-16844b987447", "address": "fa:16:3e:ca:31:df", "network": {"id": "369e3d75-3302-4816-b5f9-5da97f617152", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-855674976-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2cb049f497b14ac39a0316132f5802fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap29d0d831-eb", "ovs_interfaceid": "29d0d831-eb07-403f-bde9-16844b987447", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.147274] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Creating Snapshot of the VM instance {{(pid=61868) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 840.147559] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-e9d0b13b-22fd-45fd-89f4-5644a6fb7e68 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.155280] env[61868]: DEBUG oslo_vmware.api [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 840.155280] env[61868]: value = "task-1315390" [ 840.155280] env[61868]: _type = "Task" [ 840.155280] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.165221] env[61868]: DEBUG oslo_vmware.api [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315390, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.309540] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315387, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.317865] env[61868]: DEBUG oslo_vmware.api [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315388, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.330254] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]528671b8-5a8d-9347-11c8-61f03aabf847, 'name': SearchDatastore_Task, 'duration_secs': 0.014883} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.331046] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-51c87bbb-d2b3-49ca-ab74-c5362b8f34e7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.336102] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Waiting for the task: (returnval){ [ 840.336102] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52b82431-fa18-2c57-6c87-9d888e9ca977" [ 840.336102] env[61868]: _type = "Task" [ 840.336102] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.343750] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52b82431-fa18-2c57-6c87-9d888e9ca977, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.382969] env[61868]: DEBUG nova.compute.utils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 840.384543] env[61868]: DEBUG nova.compute.manager [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 840.386138] env[61868]: DEBUG nova.network.neutron [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 840.398949] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': task-1315389, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.441739] env[61868]: DEBUG nova.policy [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '919b1476fe6d4aa7b4663165492ea564', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8cac586898424ba3a8c64d9156a68f86', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 840.544045] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Acquiring lock "refresh_cache-52a5afc6-4e79-436c-bc94-b61ca9fb860c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.544045] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Acquired lock "refresh_cache-52a5afc6-4e79-436c-bc94-b61ca9fb860c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.544045] env[61868]: DEBUG nova.network.neutron [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 840.559335] env[61868]: DEBUG oslo_concurrency.lockutils [req-23fdc9d0-fa5c-4bfc-92db-f8d1f9d0d5a9 req-dcd0dcc2-f2c0-480f-9613-da97419bfb02 service nova] Releasing lock "refresh_cache-7788f9f8-c08e-46f8-b204-070215223436" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.668695] env[61868]: DEBUG oslo_vmware.api [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315390, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.687189] env[61868]: DEBUG nova.network.neutron [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Successfully created port: 849a9665-e0b7-4a35-a860-0ff32923cb6c {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 840.817673] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315387, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.569811} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.821532] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1/3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 840.821905] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 840.822415] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9bc824f3-9136-41f3-85fe-8465cae060d0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.830020] env[61868]: DEBUG oslo_vmware.api [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315388, 'name': ReconfigVM_Task, 'duration_secs': 0.634961} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.832726] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Reconfigured VM instance instance-00000038 to attach disk [datastore2] 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02/02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 840.833892] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Waiting for the task: (returnval){ [ 840.833892] env[61868]: value = "task-1315391" [ 840.833892] env[61868]: _type = "Task" [ 840.833892] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.833892] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-09e042d5-b608-4cb5-8664-9898051a5a1a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.856961] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315391, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.862534] env[61868]: DEBUG oslo_vmware.api [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Waiting for the task: (returnval){ [ 840.862534] env[61868]: value = "task-1315392" [ 840.862534] env[61868]: _type = "Task" [ 840.862534] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.862851] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52b82431-fa18-2c57-6c87-9d888e9ca977, 'name': SearchDatastore_Task, 'duration_secs': 0.012763} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.863208] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.863465] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 7788f9f8-c08e-46f8-b204-070215223436/7788f9f8-c08e-46f8-b204-070215223436.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 840.867271] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2a88775b-f8b1-407c-acf0-04cba26ba344 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.874536] env[61868]: DEBUG oslo_vmware.api [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315392, 'name': Rename_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.875903] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Waiting for the task: (returnval){ [ 840.875903] env[61868]: value = "task-1315393" [ 840.875903] env[61868]: _type = "Task" [ 840.875903] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.884907] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': task-1315393, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.891073] env[61868]: DEBUG nova.compute.manager [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 840.904807] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': task-1315389, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.918639] env[61868]: DEBUG nova.compute.manager [req-13ad6151-d850-4d0b-9b1c-fae4b41e8996 req-39a184b9-7687-45ea-9bd5-9d13ccfd3104 service nova] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Received event network-vif-plugged-de6238cd-1354-4052-9aff-0d28691de0c8 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 840.918903] env[61868]: DEBUG oslo_concurrency.lockutils [req-13ad6151-d850-4d0b-9b1c-fae4b41e8996 req-39a184b9-7687-45ea-9bd5-9d13ccfd3104 service nova] Acquiring lock "52a5afc6-4e79-436c-bc94-b61ca9fb860c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.919442] env[61868]: DEBUG oslo_concurrency.lockutils [req-13ad6151-d850-4d0b-9b1c-fae4b41e8996 req-39a184b9-7687-45ea-9bd5-9d13ccfd3104 service nova] Lock "52a5afc6-4e79-436c-bc94-b61ca9fb860c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.919785] env[61868]: DEBUG oslo_concurrency.lockutils [req-13ad6151-d850-4d0b-9b1c-fae4b41e8996 req-39a184b9-7687-45ea-9bd5-9d13ccfd3104 service nova] Lock "52a5afc6-4e79-436c-bc94-b61ca9fb860c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.919996] env[61868]: DEBUG nova.compute.manager [req-13ad6151-d850-4d0b-9b1c-fae4b41e8996 req-39a184b9-7687-45ea-9bd5-9d13ccfd3104 service nova] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] No waiting events found dispatching network-vif-plugged-de6238cd-1354-4052-9aff-0d28691de0c8 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 840.920315] env[61868]: WARNING nova.compute.manager [req-13ad6151-d850-4d0b-9b1c-fae4b41e8996 req-39a184b9-7687-45ea-9bd5-9d13ccfd3104 service nova] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Received unexpected event network-vif-plugged-de6238cd-1354-4052-9aff-0d28691de0c8 for instance with vm_state building and task_state spawning. [ 840.920576] env[61868]: DEBUG nova.compute.manager [req-13ad6151-d850-4d0b-9b1c-fae4b41e8996 req-39a184b9-7687-45ea-9bd5-9d13ccfd3104 service nova] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Received event network-changed-de6238cd-1354-4052-9aff-0d28691de0c8 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 840.920797] env[61868]: DEBUG nova.compute.manager [req-13ad6151-d850-4d0b-9b1c-fae4b41e8996 req-39a184b9-7687-45ea-9bd5-9d13ccfd3104 service nova] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Refreshing instance network info cache due to event network-changed-de6238cd-1354-4052-9aff-0d28691de0c8. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 840.921223] env[61868]: DEBUG oslo_concurrency.lockutils [req-13ad6151-d850-4d0b-9b1c-fae4b41e8996 req-39a184b9-7687-45ea-9bd5-9d13ccfd3104 service nova] Acquiring lock "refresh_cache-52a5afc6-4e79-436c-bc94-b61ca9fb860c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.074885] env[61868]: DEBUG nova.network.neutron [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 841.167884] env[61868]: DEBUG oslo_vmware.api [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315390, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.218885] env[61868]: DEBUG nova.network.neutron [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Updating instance_info_cache with network_info: [{"id": "de6238cd-1354-4052-9aff-0d28691de0c8", "address": "fa:16:3e:e0:aa:0c", "network": {"id": "57a3a2e2-7168-4e6c-b8e7-d4904576ffea", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1599066480-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8cac586898424ba3a8c64d9156a68f86", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1430a695-49fb-4905-bc38-db9b869a1a9d", "external-id": "nsx-vlan-transportzone-297", "segmentation_id": 297, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde6238cd-13", "ovs_interfaceid": "de6238cd-1354-4052-9aff-0d28691de0c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.224661] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e27f74e-124e-4a5d-869d-73c58ffdb07b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.232981] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9665f73-4b56-4b42-a0bf-457c7a9bcfe4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.264865] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec92ac75-c758-454b-9a9d-04a887b8b186 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.273809] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a473f3ea-8751-4785-8657-1fdedf4f7c29 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.291238] env[61868]: DEBUG nova.compute.provider_tree [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 841.349019] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315391, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062754} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.349333] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 841.350290] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c6be422-add5-4ec4-aff1-e85caf165981 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.375334] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1/3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 841.380089] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3159bc36-a7a5-4dc8-9a19-995cbca8a0f4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.408552] env[61868]: DEBUG oslo_vmware.api [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315392, 'name': Rename_Task, 'duration_secs': 0.196941} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.411910] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 841.412292] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Waiting for the task: (returnval){ [ 841.412292] env[61868]: value = "task-1315394" [ 841.412292] env[61868]: _type = "Task" [ 841.412292] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.412512] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-16516f70-d774-472c-a7b6-5b7e002b82a4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.423757] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': task-1315393, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.428239] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': task-1315389, 'name': PowerOnVM_Task, 'duration_secs': 1.486454} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.437027] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 841.437027] env[61868]: INFO nova.compute.manager [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Took 9.25 seconds to spawn the instance on the hypervisor. [ 841.437027] env[61868]: DEBUG nova.compute.manager [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 841.437027] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315394, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.437027] env[61868]: DEBUG oslo_vmware.api [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Waiting for the task: (returnval){ [ 841.437027] env[61868]: value = "task-1315395" [ 841.437027] env[61868]: _type = "Task" [ 841.437027] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.437470] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-703fa371-619e-4867-a453-c18a9a847957 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.453729] env[61868]: DEBUG oslo_vmware.api [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315395, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.666352] env[61868]: DEBUG oslo_vmware.api [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315390, 'name': CreateSnapshot_Task, 'duration_secs': 1.364205} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.666627] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Created Snapshot of the VM instance {{(pid=61868) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 841.667459] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37c82211-ad0c-47ae-8ea2-f5ad62400dc2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.721318] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Releasing lock "refresh_cache-52a5afc6-4e79-436c-bc94-b61ca9fb860c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.721711] env[61868]: DEBUG nova.compute.manager [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Instance network_info: |[{"id": "de6238cd-1354-4052-9aff-0d28691de0c8", "address": "fa:16:3e:e0:aa:0c", "network": {"id": "57a3a2e2-7168-4e6c-b8e7-d4904576ffea", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1599066480-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8cac586898424ba3a8c64d9156a68f86", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1430a695-49fb-4905-bc38-db9b869a1a9d", "external-id": "nsx-vlan-transportzone-297", "segmentation_id": 297, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde6238cd-13", "ovs_interfaceid": "de6238cd-1354-4052-9aff-0d28691de0c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 841.722018] env[61868]: DEBUG oslo_concurrency.lockutils [req-13ad6151-d850-4d0b-9b1c-fae4b41e8996 req-39a184b9-7687-45ea-9bd5-9d13ccfd3104 service nova] Acquired lock "refresh_cache-52a5afc6-4e79-436c-bc94-b61ca9fb860c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.722206] env[61868]: DEBUG nova.network.neutron [req-13ad6151-d850-4d0b-9b1c-fae4b41e8996 req-39a184b9-7687-45ea-9bd5-9d13ccfd3104 service nova] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Refreshing network info cache for port de6238cd-1354-4052-9aff-0d28691de0c8 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 841.723830] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e0:aa:0c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1430a695-49fb-4905-bc38-db9b869a1a9d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'de6238cd-1354-4052-9aff-0d28691de0c8', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 841.736449] env[61868]: DEBUG oslo.service.loopingcall [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 841.740552] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 841.741231] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6a93cba4-bac1-4757-8131-ff0fc532085f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.767586] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 841.767586] env[61868]: value = "task-1315396" [ 841.767586] env[61868]: _type = "Task" [ 841.767586] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.777409] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315396, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.794182] env[61868]: DEBUG nova.scheduler.client.report [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 841.905038] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': task-1315393, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.750079} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.905038] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 7788f9f8-c08e-46f8-b204-070215223436/7788f9f8-c08e-46f8-b204-070215223436.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 841.905038] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 841.905038] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a3736546-f517-4ed9-857b-5d21c9c82f71 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.911591] env[61868]: DEBUG nova.compute.manager [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 841.916290] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Waiting for the task: (returnval){ [ 841.916290] env[61868]: value = "task-1315397" [ 841.916290] env[61868]: _type = "Task" [ 841.916290] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.927262] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315394, 'name': ReconfigVM_Task, 'duration_secs': 0.421524} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.930305] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Reconfigured VM instance instance-0000003b to attach disk [datastore2] 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1/3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 841.930951] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': task-1315397, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.932911] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-222804e3-3f36-4c73-bf5d-d4b25bdbd3c9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.952120] env[61868]: DEBUG nova.virt.hardware [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 841.952413] env[61868]: DEBUG nova.virt.hardware [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 841.952600] env[61868]: DEBUG nova.virt.hardware [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 841.952835] env[61868]: DEBUG nova.virt.hardware [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 841.952965] env[61868]: DEBUG nova.virt.hardware [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 841.953143] env[61868]: DEBUG nova.virt.hardware [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 841.953390] env[61868]: DEBUG nova.virt.hardware [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 841.953588] env[61868]: DEBUG nova.virt.hardware [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 841.953831] env[61868]: DEBUG nova.virt.hardware [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 841.954070] env[61868]: DEBUG nova.virt.hardware [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 841.954208] env[61868]: DEBUG nova.virt.hardware [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 841.955016] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Waiting for the task: (returnval){ [ 841.955016] env[61868]: value = "task-1315398" [ 841.955016] env[61868]: _type = "Task" [ 841.955016] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.960695] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-196aec7f-e1cb-4d45-ad4b-310c42112626 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.968905] env[61868]: INFO nova.compute.manager [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Took 41.91 seconds to build instance. [ 841.979575] env[61868]: DEBUG oslo_vmware.api [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315395, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.982771] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-067dbf63-a024-4c1e-9a74-1f399d2b3cf6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.992366] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315398, 'name': Rename_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.030735] env[61868]: DEBUG nova.network.neutron [req-13ad6151-d850-4d0b-9b1c-fae4b41e8996 req-39a184b9-7687-45ea-9bd5-9d13ccfd3104 service nova] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Updated VIF entry in instance network info cache for port de6238cd-1354-4052-9aff-0d28691de0c8. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 842.031250] env[61868]: DEBUG nova.network.neutron [req-13ad6151-d850-4d0b-9b1c-fae4b41e8996 req-39a184b9-7687-45ea-9bd5-9d13ccfd3104 service nova] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Updating instance_info_cache with network_info: [{"id": "de6238cd-1354-4052-9aff-0d28691de0c8", "address": "fa:16:3e:e0:aa:0c", "network": {"id": "57a3a2e2-7168-4e6c-b8e7-d4904576ffea", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1599066480-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8cac586898424ba3a8c64d9156a68f86", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1430a695-49fb-4905-bc38-db9b869a1a9d", "external-id": "nsx-vlan-transportzone-297", "segmentation_id": 297, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde6238cd-13", "ovs_interfaceid": "de6238cd-1354-4052-9aff-0d28691de0c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.186498] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Creating linked-clone VM from snapshot {{(pid=61868) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 842.187027] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-14b55012-c5b3-4ba7-a36b-ac98781ad132 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.195160] env[61868]: DEBUG oslo_vmware.api [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 842.195160] env[61868]: value = "task-1315399" [ 842.195160] env[61868]: _type = "Task" [ 842.195160] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.204229] env[61868]: DEBUG oslo_vmware.api [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315399, 'name': CloneVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.244843] env[61868]: DEBUG nova.network.neutron [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Successfully updated port: 849a9665-e0b7-4a35-a860-0ff32923cb6c {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 842.277986] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315396, 'name': CreateVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.299643] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.418s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.300429] env[61868]: DEBUG nova.compute.manager [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 842.303479] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.765s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.304910] env[61868]: INFO nova.compute.claims [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 842.431705] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': task-1315397, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.14056} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.431705] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 842.431705] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aebcd1f9-f18e-4aa4-8bcd-fc6279d5fc7b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.455020] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] 7788f9f8-c08e-46f8-b204-070215223436/7788f9f8-c08e-46f8-b204-070215223436.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 842.458434] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f113a646-7946-4bf4-bfeb-a3ef9a101540 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.479111] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Lock "0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 107.386s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.487777] env[61868]: DEBUG oslo_vmware.api [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315395, 'name': PowerOnVM_Task, 'duration_secs': 0.576931} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.489303] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 842.489547] env[61868]: DEBUG nova.compute.manager [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 842.489988] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Waiting for the task: (returnval){ [ 842.489988] env[61868]: value = "task-1315400" [ 842.489988] env[61868]: _type = "Task" [ 842.489988] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.490637] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd4f2ec4-5f37-4928-8a00-7b62f52e4936 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.498391] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315398, 'name': Rename_Task, 'duration_secs': 0.270451} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.498995] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 842.499515] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2a42d505-ed90-441b-84e9-abfa507f71f5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.510073] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': task-1315400, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.514454] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Waiting for the task: (returnval){ [ 842.514454] env[61868]: value = "task-1315401" [ 842.514454] env[61868]: _type = "Task" [ 842.514454] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.524681] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315401, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.534779] env[61868]: DEBUG oslo_concurrency.lockutils [req-13ad6151-d850-4d0b-9b1c-fae4b41e8996 req-39a184b9-7687-45ea-9bd5-9d13ccfd3104 service nova] Releasing lock "refresh_cache-52a5afc6-4e79-436c-bc94-b61ca9fb860c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.706624] env[61868]: DEBUG oslo_vmware.api [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315399, 'name': CloneVM_Task} progress is 94%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.748092] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Acquiring lock "refresh_cache-c2693a43-4ea2-4ab0-8915-2fa544780e3c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.748250] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Acquired lock "refresh_cache-c2693a43-4ea2-4ab0-8915-2fa544780e3c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.748401] env[61868]: DEBUG nova.network.neutron [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 842.781585] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315396, 'name': CreateVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.809438] env[61868]: DEBUG nova.compute.utils [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 842.814326] env[61868]: DEBUG nova.compute.manager [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 842.814617] env[61868]: DEBUG nova.network.neutron [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 843.003112] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': task-1315400, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.005445] env[61868]: DEBUG nova.compute.manager [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 843.024022] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.025492] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315401, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.031323] env[61868]: DEBUG nova.compute.manager [req-b45938ce-3115-4765-af62-134693690fcf req-cb642513-7d1c-406e-94dc-02b31d99fc5c service nova] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Received event network-vif-plugged-849a9665-e0b7-4a35-a860-0ff32923cb6c {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 843.031831] env[61868]: DEBUG oslo_concurrency.lockutils [req-b45938ce-3115-4765-af62-134693690fcf req-cb642513-7d1c-406e-94dc-02b31d99fc5c service nova] Acquiring lock "c2693a43-4ea2-4ab0-8915-2fa544780e3c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.032174] env[61868]: DEBUG oslo_concurrency.lockutils [req-b45938ce-3115-4765-af62-134693690fcf req-cb642513-7d1c-406e-94dc-02b31d99fc5c service nova] Lock "c2693a43-4ea2-4ab0-8915-2fa544780e3c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.032452] env[61868]: DEBUG oslo_concurrency.lockutils [req-b45938ce-3115-4765-af62-134693690fcf req-cb642513-7d1c-406e-94dc-02b31d99fc5c service nova] Lock "c2693a43-4ea2-4ab0-8915-2fa544780e3c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.032732] env[61868]: DEBUG nova.compute.manager [req-b45938ce-3115-4765-af62-134693690fcf req-cb642513-7d1c-406e-94dc-02b31d99fc5c service nova] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] No waiting events found dispatching network-vif-plugged-849a9665-e0b7-4a35-a860-0ff32923cb6c {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 843.033824] env[61868]: WARNING nova.compute.manager [req-b45938ce-3115-4765-af62-134693690fcf req-cb642513-7d1c-406e-94dc-02b31d99fc5c service nova] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Received unexpected event network-vif-plugged-849a9665-e0b7-4a35-a860-0ff32923cb6c for instance with vm_state building and task_state spawning. [ 843.033824] env[61868]: DEBUG nova.compute.manager [req-b45938ce-3115-4765-af62-134693690fcf req-cb642513-7d1c-406e-94dc-02b31d99fc5c service nova] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Received event network-changed-849a9665-e0b7-4a35-a860-0ff32923cb6c {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 843.033824] env[61868]: DEBUG nova.compute.manager [req-b45938ce-3115-4765-af62-134693690fcf req-cb642513-7d1c-406e-94dc-02b31d99fc5c service nova] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Refreshing instance network info cache due to event network-changed-849a9665-e0b7-4a35-a860-0ff32923cb6c. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 843.033824] env[61868]: DEBUG oslo_concurrency.lockutils [req-b45938ce-3115-4765-af62-134693690fcf req-cb642513-7d1c-406e-94dc-02b31d99fc5c service nova] Acquiring lock "refresh_cache-c2693a43-4ea2-4ab0-8915-2fa544780e3c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.047721] env[61868]: DEBUG nova.policy [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '379bc4b2a0e34a779eb064dd18f02a08', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7471600a510e4feeb0c819d0cca5eb48', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 843.208596] env[61868]: DEBUG oslo_vmware.api [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315399, 'name': CloneVM_Task} progress is 94%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.280251] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315396, 'name': CreateVM_Task, 'duration_secs': 1.126257} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.280397] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 843.281091] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.281266] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.281686] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 843.282413] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d390ad94-77e8-4cb5-9bdb-9c231eed6a1b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.286526] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Waiting for the task: (returnval){ [ 843.286526] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52c09eb1-ccba-997c-b7c8-629ec24522ba" [ 843.286526] env[61868]: _type = "Task" [ 843.286526] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.295357] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52c09eb1-ccba-997c-b7c8-629ec24522ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.301881] env[61868]: DEBUG nova.network.neutron [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 843.317759] env[61868]: DEBUG nova.compute.manager [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 843.381220] env[61868]: DEBUG nova.network.neutron [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Successfully created port: 29d18263-e917-46dc-8b91-33365c0d3481 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 843.507461] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': task-1315400, 'name': ReconfigVM_Task, 'duration_secs': 0.539033} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.507643] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Reconfigured VM instance instance-0000003d to attach disk [datastore2] 7788f9f8-c08e-46f8-b204-070215223436/7788f9f8-c08e-46f8-b204-070215223436.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 843.508565] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-68cc9247-c782-4c8b-8a77-35f8aca8cc3a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.520574] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Waiting for the task: (returnval){ [ 843.520574] env[61868]: value = "task-1315402" [ 843.520574] env[61868]: _type = "Task" [ 843.520574] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.528707] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315401, 'name': PowerOnVM_Task, 'duration_secs': 0.650927} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.529553] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.530660] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 843.530660] env[61868]: INFO nova.compute.manager [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Took 8.96 seconds to spawn the instance on the hypervisor. [ 843.530834] env[61868]: DEBUG nova.compute.manager [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 843.531518] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e89cf70-197e-4496-8427-70de491e8da4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.537169] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': task-1315402, 'name': Rename_Task} progress is 10%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.571659] env[61868]: DEBUG nova.network.neutron [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Updating instance_info_cache with network_info: [{"id": "849a9665-e0b7-4a35-a860-0ff32923cb6c", "address": "fa:16:3e:49:f0:08", "network": {"id": "57a3a2e2-7168-4e6c-b8e7-d4904576ffea", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1599066480-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8cac586898424ba3a8c64d9156a68f86", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1430a695-49fb-4905-bc38-db9b869a1a9d", "external-id": "nsx-vlan-transportzone-297", "segmentation_id": 297, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap849a9665-e0", "ovs_interfaceid": "849a9665-e0b7-4a35-a860-0ff32923cb6c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.682653] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cfb29b5-4b94-4298-8e59-1abae7d18ddc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.691363] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6159350-5a27-4a03-b4b3-c29311f7ea50 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.733770] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b043ef1d-e6c9-451b-b682-a9d802d40811 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.743110] env[61868]: DEBUG oslo_vmware.api [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315399, 'name': CloneVM_Task} progress is 94%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.747698] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bed6788-df2d-4cd3-9cbf-0fa92ebb75cf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.762757] env[61868]: DEBUG nova.compute.provider_tree [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 843.796230] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52c09eb1-ccba-997c-b7c8-629ec24522ba, 'name': SearchDatastore_Task, 'duration_secs': 0.018833} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.796525] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.796767] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 843.796998] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.797159] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.797338] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 843.797598] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-44c659e9-b530-4562-88f0-0751b8fbfb88 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.808337] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 843.808516] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 843.809229] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a95e2cf7-5abd-4dd9-ae87-ad7b63b62258 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.814833] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Waiting for the task: (returnval){ [ 843.814833] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]525911c4-9d2c-c61f-9e3e-fef9ef86bc37" [ 843.814833] env[61868]: _type = "Task" [ 843.814833] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.825323] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]525911c4-9d2c-c61f-9e3e-fef9ef86bc37, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.034035] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': task-1315402, 'name': Rename_Task, 'duration_secs': 0.184574} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.034035] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 844.034035] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-40a9998e-c8ca-4121-8727-ab32f35a5f23 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.040536] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Waiting for the task: (returnval){ [ 844.040536] env[61868]: value = "task-1315403" [ 844.040536] env[61868]: _type = "Task" [ 844.040536] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.053832] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': task-1315403, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.061534] env[61868]: INFO nova.compute.manager [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Took 27.70 seconds to build instance. [ 844.076937] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Releasing lock "refresh_cache-c2693a43-4ea2-4ab0-8915-2fa544780e3c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.077290] env[61868]: DEBUG nova.compute.manager [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Instance network_info: |[{"id": "849a9665-e0b7-4a35-a860-0ff32923cb6c", "address": "fa:16:3e:49:f0:08", "network": {"id": "57a3a2e2-7168-4e6c-b8e7-d4904576ffea", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1599066480-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8cac586898424ba3a8c64d9156a68f86", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1430a695-49fb-4905-bc38-db9b869a1a9d", "external-id": "nsx-vlan-transportzone-297", "segmentation_id": 297, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap849a9665-e0", "ovs_interfaceid": "849a9665-e0b7-4a35-a860-0ff32923cb6c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 844.080135] env[61868]: DEBUG oslo_concurrency.lockutils [req-b45938ce-3115-4765-af62-134693690fcf req-cb642513-7d1c-406e-94dc-02b31d99fc5c service nova] Acquired lock "refresh_cache-c2693a43-4ea2-4ab0-8915-2fa544780e3c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.080135] env[61868]: DEBUG nova.network.neutron [req-b45938ce-3115-4765-af62-134693690fcf req-cb642513-7d1c-406e-94dc-02b31d99fc5c service nova] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Refreshing network info cache for port 849a9665-e0b7-4a35-a860-0ff32923cb6c {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 844.080135] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:49:f0:08', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1430a695-49fb-4905-bc38-db9b869a1a9d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '849a9665-e0b7-4a35-a860-0ff32923cb6c', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 844.086665] env[61868]: DEBUG oslo.service.loopingcall [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 844.087391] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 844.087618] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-eb887cbf-6eb0-438e-8990-138febff58bb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.109035] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 844.109035] env[61868]: value = "task-1315404" [ 844.109035] env[61868]: _type = "Task" [ 844.109035] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.118554] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315404, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.211895] env[61868]: DEBUG oslo_vmware.api [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315399, 'name': CloneVM_Task, 'duration_secs': 1.972629} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.211895] env[61868]: INFO nova.virt.vmwareapi.vmops [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Created linked-clone VM from snapshot [ 844.211895] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfe09dfa-6458-43fd-8027-87603625b64f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.217566] env[61868]: DEBUG nova.virt.vmwareapi.images [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Uploading image 577e8014-2c7d-4375-ad66-8f0ffb82b743 {{(pid=61868) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 844.238384] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Acquiring lock "02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.238624] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Lock "02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.238827] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Acquiring lock "02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.239016] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Lock "02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.239181] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Lock "02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.244927] env[61868]: DEBUG oslo_vmware.rw_handles [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 844.244927] env[61868]: value = "vm-281518" [ 844.244927] env[61868]: _type = "VirtualMachine" [ 844.244927] env[61868]: }. {{(pid=61868) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 844.245234] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-05397a6d-0c65-42bc-9d7d-335b559f0197 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.247365] env[61868]: INFO nova.compute.manager [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Terminating instance [ 844.249638] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Acquiring lock "refresh_cache-02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.249822] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Acquired lock "refresh_cache-02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.250008] env[61868]: DEBUG nova.network.neutron [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 844.254124] env[61868]: DEBUG oslo_vmware.rw_handles [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lease: (returnval){ [ 844.254124] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52ef5254-85be-3042-01d9-1e2cad0ea9a6" [ 844.254124] env[61868]: _type = "HttpNfcLease" [ 844.254124] env[61868]: } obtained for exporting VM: (result){ [ 844.254124] env[61868]: value = "vm-281518" [ 844.254124] env[61868]: _type = "VirtualMachine" [ 844.254124] env[61868]: }. {{(pid=61868) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 844.254420] env[61868]: DEBUG oslo_vmware.api [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the lease: (returnval){ [ 844.254420] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52ef5254-85be-3042-01d9-1e2cad0ea9a6" [ 844.254420] env[61868]: _type = "HttpNfcLease" [ 844.254420] env[61868]: } to be ready. {{(pid=61868) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 844.261277] env[61868]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 844.261277] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52ef5254-85be-3042-01d9-1e2cad0ea9a6" [ 844.261277] env[61868]: _type = "HttpNfcLease" [ 844.261277] env[61868]: } is initializing. {{(pid=61868) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 844.266264] env[61868]: DEBUG nova.scheduler.client.report [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 844.326061] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]525911c4-9d2c-c61f-9e3e-fef9ef86bc37, 'name': SearchDatastore_Task, 'duration_secs': 0.012861} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.327245] env[61868]: DEBUG nova.compute.manager [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 844.329891] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0bcf341d-d565-4a8c-a3f9-75780e430919 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.335213] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Waiting for the task: (returnval){ [ 844.335213] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52401e03-efb8-cdf7-8bfb-bebe92a87948" [ 844.335213] env[61868]: _type = "Task" [ 844.335213] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.342995] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52401e03-efb8-cdf7-8bfb-bebe92a87948, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.350694] env[61868]: DEBUG nova.virt.hardware [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 844.350961] env[61868]: DEBUG nova.virt.hardware [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 844.351152] env[61868]: DEBUG nova.virt.hardware [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 844.351345] env[61868]: DEBUG nova.virt.hardware [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 844.351546] env[61868]: DEBUG nova.virt.hardware [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 844.351677] env[61868]: DEBUG nova.virt.hardware [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 844.351902] env[61868]: DEBUG nova.virt.hardware [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 844.352076] env[61868]: DEBUG nova.virt.hardware [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 844.352243] env[61868]: DEBUG nova.virt.hardware [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 844.352403] env[61868]: DEBUG nova.virt.hardware [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 844.352611] env[61868]: DEBUG nova.virt.hardware [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 844.353752] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1d30b51-1122-43c5-a0fc-b5eb28dca9de {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.361560] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ce3f286-ae32-4258-852a-ebce8d5719d3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.550995] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': task-1315403, 'name': PowerOnVM_Task} progress is 90%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.563808] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Lock "3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.452s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.620297] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315404, 'name': CreateVM_Task, 'duration_secs': 0.351841} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.620531] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 844.621278] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.621449] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.621801] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 844.622083] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9fb07d6e-239a-45ea-9f66-94ec5cc086e4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.627245] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Waiting for the task: (returnval){ [ 844.627245] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5261eb22-e84b-38de-51c9-35b2aa134f89" [ 844.627245] env[61868]: _type = "Task" [ 844.627245] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.635250] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5261eb22-e84b-38de-51c9-35b2aa134f89, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.761833] env[61868]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 844.761833] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52ef5254-85be-3042-01d9-1e2cad0ea9a6" [ 844.761833] env[61868]: _type = "HttpNfcLease" [ 844.761833] env[61868]: } is ready. {{(pid=61868) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 844.762147] env[61868]: DEBUG oslo_vmware.rw_handles [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 844.762147] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52ef5254-85be-3042-01d9-1e2cad0ea9a6" [ 844.762147] env[61868]: _type = "HttpNfcLease" [ 844.762147] env[61868]: }. {{(pid=61868) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 844.762879] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e859d07d-f5e5-4364-bb3f-683165531457 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.777428] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.474s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.778040] env[61868]: DEBUG nova.compute.manager [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 844.780974] env[61868]: DEBUG oslo_vmware.rw_handles [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5276e24e-54b4-9464-2978-4c5f0e74aec0/disk-0.vmdk from lease info. {{(pid=61868) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 844.781169] env[61868]: DEBUG oslo_vmware.rw_handles [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5276e24e-54b4-9464-2978-4c5f0e74aec0/disk-0.vmdk for reading. {{(pid=61868) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 844.787703] env[61868]: DEBUG nova.network.neutron [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 844.792274] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.495s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.794429] env[61868]: INFO nova.compute.claims [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 844.867567] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52401e03-efb8-cdf7-8bfb-bebe92a87948, 'name': SearchDatastore_Task, 'duration_secs': 0.010796} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.870068] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.870408] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 52a5afc6-4e79-436c-bc94-b61ca9fb860c/52a5afc6-4e79-436c-bc94-b61ca9fb860c.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 844.870838] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-07a748dc-8945-4d76-aca8-902dcdd08e46 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.879567] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Waiting for the task: (returnval){ [ 844.879567] env[61868]: value = "task-1315406" [ 844.879567] env[61868]: _type = "Task" [ 844.879567] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.888928] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315406, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.916075] env[61868]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-5a395e5f-e663-4cfe-acb1-c269dc095bc9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.963863] env[61868]: DEBUG nova.network.neutron [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.986056] env[61868]: DEBUG nova.network.neutron [req-b45938ce-3115-4765-af62-134693690fcf req-cb642513-7d1c-406e-94dc-02b31d99fc5c service nova] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Updated VIF entry in instance network info cache for port 849a9665-e0b7-4a35-a860-0ff32923cb6c. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 844.986497] env[61868]: DEBUG nova.network.neutron [req-b45938ce-3115-4765-af62-134693690fcf req-cb642513-7d1c-406e-94dc-02b31d99fc5c service nova] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Updating instance_info_cache with network_info: [{"id": "849a9665-e0b7-4a35-a860-0ff32923cb6c", "address": "fa:16:3e:49:f0:08", "network": {"id": "57a3a2e2-7168-4e6c-b8e7-d4904576ffea", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1599066480-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8cac586898424ba3a8c64d9156a68f86", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1430a695-49fb-4905-bc38-db9b869a1a9d", "external-id": "nsx-vlan-transportzone-297", "segmentation_id": 297, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap849a9665-e0", "ovs_interfaceid": "849a9665-e0b7-4a35-a860-0ff32923cb6c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.049895] env[61868]: DEBUG oslo_vmware.api [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': task-1315403, 'name': PowerOnVM_Task, 'duration_secs': 0.610593} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.050172] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 845.050367] env[61868]: INFO nova.compute.manager [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Took 8.12 seconds to spawn the instance on the hypervisor. [ 845.050541] env[61868]: DEBUG nova.compute.manager [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 845.051314] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70186ef8-e710-4564-9b2d-fc89fa35871d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.069550] env[61868]: DEBUG nova.compute.manager [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 845.076703] env[61868]: DEBUG nova.compute.manager [req-f8b586b3-ba73-4c26-a8bf-9c3b5a0b898e req-98ab5482-6557-4a5a-bd03-895aafd98420 service nova] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Received event network-vif-plugged-29d18263-e917-46dc-8b91-33365c0d3481 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 845.076908] env[61868]: DEBUG oslo_concurrency.lockutils [req-f8b586b3-ba73-4c26-a8bf-9c3b5a0b898e req-98ab5482-6557-4a5a-bd03-895aafd98420 service nova] Acquiring lock "409cad1a-946d-4c58-aa57-1c0bf97fe63f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.077120] env[61868]: DEBUG oslo_concurrency.lockutils [req-f8b586b3-ba73-4c26-a8bf-9c3b5a0b898e req-98ab5482-6557-4a5a-bd03-895aafd98420 service nova] Lock "409cad1a-946d-4c58-aa57-1c0bf97fe63f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.077280] env[61868]: DEBUG oslo_concurrency.lockutils [req-f8b586b3-ba73-4c26-a8bf-9c3b5a0b898e req-98ab5482-6557-4a5a-bd03-895aafd98420 service nova] Lock "409cad1a-946d-4c58-aa57-1c0bf97fe63f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.077442] env[61868]: DEBUG nova.compute.manager [req-f8b586b3-ba73-4c26-a8bf-9c3b5a0b898e req-98ab5482-6557-4a5a-bd03-895aafd98420 service nova] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] No waiting events found dispatching network-vif-plugged-29d18263-e917-46dc-8b91-33365c0d3481 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 845.077659] env[61868]: WARNING nova.compute.manager [req-f8b586b3-ba73-4c26-a8bf-9c3b5a0b898e req-98ab5482-6557-4a5a-bd03-895aafd98420 service nova] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Received unexpected event network-vif-plugged-29d18263-e917-46dc-8b91-33365c0d3481 for instance with vm_state building and task_state spawning. [ 845.111335] env[61868]: DEBUG nova.network.neutron [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Successfully updated port: 29d18263-e917-46dc-8b91-33365c0d3481 {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 845.139268] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5261eb22-e84b-38de-51c9-35b2aa134f89, 'name': SearchDatastore_Task, 'duration_secs': 0.013204} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.139586] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.139830] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 845.140116] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.140346] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.140445] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 845.140649] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1725d794-22c6-4294-8662-8697bd036f0c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.154034] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 845.154241] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 845.154982] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-28c1ee1a-bab7-4f64-8f9f-6f3de4dec39c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.159950] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Waiting for the task: (returnval){ [ 845.159950] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e8650c-d5b4-157d-3d10-40b949759250" [ 845.159950] env[61868]: _type = "Task" [ 845.159950] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.174267] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e8650c-d5b4-157d-3d10-40b949759250, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.291981] env[61868]: DEBUG nova.compute.utils [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 845.295543] env[61868]: DEBUG nova.compute.manager [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 845.295543] env[61868]: DEBUG nova.network.neutron [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 845.358733] env[61868]: DEBUG nova.policy [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e02ffdc140274733a081fd53c4acc202', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a60c3c20950e4156b64c3b4c61b9f0f2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 845.394166] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315406, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.466727] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Releasing lock "refresh_cache-02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.467340] env[61868]: DEBUG nova.compute.manager [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 845.467551] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 845.468564] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7da61481-bab7-469b-9bcd-1ec787a4ba0e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.480803] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 845.481760] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-96d0e02a-2053-485c-bef6-365ec09464de {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.490321] env[61868]: DEBUG oslo_concurrency.lockutils [req-b45938ce-3115-4765-af62-134693690fcf req-cb642513-7d1c-406e-94dc-02b31d99fc5c service nova] Releasing lock "refresh_cache-c2693a43-4ea2-4ab0-8915-2fa544780e3c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.492873] env[61868]: DEBUG oslo_vmware.api [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Waiting for the task: (returnval){ [ 845.492873] env[61868]: value = "task-1315407" [ 845.492873] env[61868]: _type = "Task" [ 845.492873] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.504194] env[61868]: DEBUG oslo_vmware.api [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315407, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.573241] env[61868]: INFO nova.compute.manager [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Took 24.58 seconds to build instance. [ 845.588884] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.615355] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Acquiring lock "refresh_cache-409cad1a-946d-4c58-aa57-1c0bf97fe63f" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.615633] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Acquired lock "refresh_cache-409cad1a-946d-4c58-aa57-1c0bf97fe63f" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.615891] env[61868]: DEBUG nova.network.neutron [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 845.673224] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e8650c-d5b4-157d-3d10-40b949759250, 'name': SearchDatastore_Task, 'duration_secs': 0.021433} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.678653] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c34603e-7348-4359-b7a7-4179c71c9b58 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.683985] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Waiting for the task: (returnval){ [ 845.683985] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52715036-f2d8-039d-a25b-a50418699e2a" [ 845.683985] env[61868]: _type = "Task" [ 845.683985] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.692359] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52715036-f2d8-039d-a25b-a50418699e2a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.697506] env[61868]: DEBUG nova.network.neutron [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Successfully created port: 7b94f787-9d7e-4edf-92b0-44d37a7e93ec {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 845.746535] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aea8c5e2-c1b8-4c5c-b163-14cb1796a1cf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.755009] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a48657e9-b0a5-42a7-83be-91e0506733cc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.787765] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f32e71e-0973-49a8-919b-c823744ac5a5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.796694] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2416af09-bcd5-4589-a8d7-748210c0bc19 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.801086] env[61868]: DEBUG nova.compute.manager [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 845.817252] env[61868]: DEBUG nova.compute.provider_tree [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 845.890992] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315406, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.71721} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.891715] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 52a5afc6-4e79-436c-bc94-b61ca9fb860c/52a5afc6-4e79-436c-bc94-b61ca9fb860c.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 845.892134] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 845.892528] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-080c713d-e52a-407b-9013-6b048addeffa {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.899837] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Waiting for the task: (returnval){ [ 845.899837] env[61868]: value = "task-1315408" [ 845.899837] env[61868]: _type = "Task" [ 845.899837] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.909159] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315408, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.003592] env[61868]: DEBUG oslo_vmware.api [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315407, 'name': PowerOffVM_Task, 'duration_secs': 0.195209} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.003721] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 846.003892] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 846.004299] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-90f26d73-c3d1-4991-b5ee-c4c28b2d5efc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.029992] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 846.030209] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Deleting contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 846.030399] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Deleting the datastore file [datastore2] 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 846.030783] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d8c78f13-81e5-42a7-9956-32e7b309a2a4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.037703] env[61868]: DEBUG oslo_vmware.api [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Waiting for the task: (returnval){ [ 846.037703] env[61868]: value = "task-1315410" [ 846.037703] env[61868]: _type = "Task" [ 846.037703] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.046175] env[61868]: DEBUG oslo_vmware.api [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315410, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.076240] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03292ce2-b245-4d93-9068-70a38ed7d845 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Lock "7788f9f8-c08e-46f8-b204-070215223436" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.948s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.151823] env[61868]: DEBUG nova.network.neutron [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 846.197273] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52715036-f2d8-039d-a25b-a50418699e2a, 'name': SearchDatastore_Task, 'duration_secs': 0.008618} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.197959] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.198244] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] c2693a43-4ea2-4ab0-8915-2fa544780e3c/c2693a43-4ea2-4ab0-8915-2fa544780e3c.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 846.198530] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5568cea1-5907-4980-8f31-b1b4065e7983 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.205277] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Waiting for the task: (returnval){ [ 846.205277] env[61868]: value = "task-1315411" [ 846.205277] env[61868]: _type = "Task" [ 846.205277] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.214223] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315411, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.310531] env[61868]: DEBUG nova.network.neutron [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Updating instance_info_cache with network_info: [{"id": "29d18263-e917-46dc-8b91-33365c0d3481", "address": "fa:16:3e:c6:07:cd", "network": {"id": "0fa2f8e1-ee57-450f-9d06-9a056566585b", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-729551994-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7471600a510e4feeb0c819d0cca5eb48", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap29d18263-e9", "ovs_interfaceid": "29d18263-e917-46dc-8b91-33365c0d3481", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.318851] env[61868]: DEBUG nova.scheduler.client.report [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 846.411062] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315408, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069124} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.411394] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 846.412292] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21af19d9-91dc-431f-a231-a1d572c26d8b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.437344] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] 52a5afc6-4e79-436c-bc94-b61ca9fb860c/52a5afc6-4e79-436c-bc94-b61ca9fb860c.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 846.437703] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e4169fce-3e20-4bc0-828e-e817cb63e41c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.462065] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Waiting for the task: (returnval){ [ 846.462065] env[61868]: value = "task-1315412" [ 846.462065] env[61868]: _type = "Task" [ 846.462065] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.472546] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315412, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.548728] env[61868]: DEBUG oslo_vmware.api [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315410, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.098583} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.549105] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 846.549381] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Deleted contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 846.549605] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 846.549826] env[61868]: INFO nova.compute.manager [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Took 1.08 seconds to destroy the instance on the hypervisor. [ 846.550204] env[61868]: DEBUG oslo.service.loopingcall [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 846.550464] env[61868]: DEBUG nova.compute.manager [-] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 846.550566] env[61868]: DEBUG nova.network.neutron [-] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 846.569084] env[61868]: DEBUG nova.network.neutron [-] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 846.580842] env[61868]: DEBUG nova.compute.manager [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 846.647752] env[61868]: DEBUG oslo_concurrency.lockutils [None req-fd684283-1baf-4fb4-90ff-bd8128b2a028 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquiring lock "0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.647925] env[61868]: DEBUG oslo_concurrency.lockutils [None req-fd684283-1baf-4fb4-90ff-bd8128b2a028 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Lock "0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.648127] env[61868]: DEBUG oslo_concurrency.lockutils [None req-fd684283-1baf-4fb4-90ff-bd8128b2a028 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquiring lock "0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.648332] env[61868]: DEBUG oslo_concurrency.lockutils [None req-fd684283-1baf-4fb4-90ff-bd8128b2a028 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Lock "0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.648528] env[61868]: DEBUG oslo_concurrency.lockutils [None req-fd684283-1baf-4fb4-90ff-bd8128b2a028 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Lock "0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.651464] env[61868]: INFO nova.compute.manager [None req-fd684283-1baf-4fb4-90ff-bd8128b2a028 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Terminating instance [ 846.653835] env[61868]: DEBUG nova.compute.manager [None req-fd684283-1baf-4fb4-90ff-bd8128b2a028 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 846.654109] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-fd684283-1baf-4fb4-90ff-bd8128b2a028 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 846.655020] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-590021f7-495b-4097-808a-66b3998ba212 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.664677] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd684283-1baf-4fb4-90ff-bd8128b2a028 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 846.664973] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4509b05e-f0e9-4d1a-8803-a6c4abc93945 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.673753] env[61868]: DEBUG oslo_vmware.api [None req-fd684283-1baf-4fb4-90ff-bd8128b2a028 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Waiting for the task: (returnval){ [ 846.673753] env[61868]: value = "task-1315413" [ 846.673753] env[61868]: _type = "Task" [ 846.673753] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.682184] env[61868]: DEBUG oslo_vmware.api [None req-fd684283-1baf-4fb4-90ff-bd8128b2a028 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': task-1315413, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.714870] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315411, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.485435} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.715156] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] c2693a43-4ea2-4ab0-8915-2fa544780e3c/c2693a43-4ea2-4ab0-8915-2fa544780e3c.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 846.715364] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 846.715624] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0b6a940e-3c79-4f17-8e74-94136f80594f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.722305] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Waiting for the task: (returnval){ [ 846.722305] env[61868]: value = "task-1315414" [ 846.722305] env[61868]: _type = "Task" [ 846.722305] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.730781] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315414, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.813098] env[61868]: DEBUG nova.compute.manager [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 846.816497] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Releasing lock "refresh_cache-409cad1a-946d-4c58-aa57-1c0bf97fe63f" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.816884] env[61868]: DEBUG nova.compute.manager [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Instance network_info: |[{"id": "29d18263-e917-46dc-8b91-33365c0d3481", "address": "fa:16:3e:c6:07:cd", "network": {"id": "0fa2f8e1-ee57-450f-9d06-9a056566585b", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-729551994-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7471600a510e4feeb0c819d0cca5eb48", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap29d18263-e9", "ovs_interfaceid": "29d18263-e917-46dc-8b91-33365c0d3481", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 846.817424] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c6:07:cd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '184687d6-125a-4b58-bb5b-fdb404088eda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '29d18263-e917-46dc-8b91-33365c0d3481', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 846.825624] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Creating folder: Project (7471600a510e4feeb0c819d0cca5eb48). Parent ref: group-v281478. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 846.826484] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.034s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.827022] env[61868]: DEBUG nova.compute.manager [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 846.829748] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4512525b-840a-47be-8dae-1291863a4ad5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.831937] env[61868]: DEBUG oslo_concurrency.lockutils [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.745s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.833460] env[61868]: INFO nova.compute.claims [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 846.847142] env[61868]: DEBUG nova.virt.hardware [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 846.847429] env[61868]: DEBUG nova.virt.hardware [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 846.847583] env[61868]: DEBUG nova.virt.hardware [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 846.847779] env[61868]: DEBUG nova.virt.hardware [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 846.847975] env[61868]: DEBUG nova.virt.hardware [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 846.848155] env[61868]: DEBUG nova.virt.hardware [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 846.848336] env[61868]: DEBUG nova.virt.hardware [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 846.848554] env[61868]: DEBUG nova.virt.hardware [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 846.848716] env[61868]: DEBUG nova.virt.hardware [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 846.848840] env[61868]: DEBUG nova.virt.hardware [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 846.849019] env[61868]: DEBUG nova.virt.hardware [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 846.849473] env[61868]: DEBUG oslo_concurrency.lockutils [None req-05a0085d-ae8b-4e73-9106-330b108967d5 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquiring lock "7788f9f8-c08e-46f8-b204-070215223436" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.849702] env[61868]: DEBUG oslo_concurrency.lockutils [None req-05a0085d-ae8b-4e73-9106-330b108967d5 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Lock "7788f9f8-c08e-46f8-b204-070215223436" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.849958] env[61868]: DEBUG oslo_concurrency.lockutils [None req-05a0085d-ae8b-4e73-9106-330b108967d5 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquiring lock "7788f9f8-c08e-46f8-b204-070215223436-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.850084] env[61868]: DEBUG oslo_concurrency.lockutils [None req-05a0085d-ae8b-4e73-9106-330b108967d5 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Lock "7788f9f8-c08e-46f8-b204-070215223436-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.850258] env[61868]: DEBUG oslo_concurrency.lockutils [None req-05a0085d-ae8b-4e73-9106-330b108967d5 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Lock "7788f9f8-c08e-46f8-b204-070215223436-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.853727] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8dba8dd-d139-42a8-891d-14b279d3a44e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.856781] env[61868]: INFO nova.compute.manager [None req-05a0085d-ae8b-4e73-9106-330b108967d5 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Terminating instance [ 846.859581] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Created folder: Project (7471600a510e4feeb0c819d0cca5eb48) in parent group-v281478. [ 846.860056] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Creating folder: Instances. Parent ref: group-v281520. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 846.860400] env[61868]: DEBUG nova.compute.manager [None req-05a0085d-ae8b-4e73-9106-330b108967d5 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 846.860600] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-05a0085d-ae8b-4e73-9106-330b108967d5 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 846.861239] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e0c3c880-51f7-43fd-bb01-c489e0ca7e1b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.863467] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-958c4a7d-b3c2-4a8e-a6c3-694f52c56ae4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.871150] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cf474be-ecf2-4807-a756-48b9601a636c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.877639] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-05a0085d-ae8b-4e73-9106-330b108967d5 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 846.879820] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d4a83240-953d-4b76-80b3-40f92cc0cc39 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.881391] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Created folder: Instances in parent group-v281520. [ 846.881678] env[61868]: DEBUG oslo.service.loopingcall [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 846.890222] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 846.891133] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cef8847c-2fef-4940-9ff4-b8be29cbae4b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.907023] env[61868]: DEBUG oslo_vmware.api [None req-05a0085d-ae8b-4e73-9106-330b108967d5 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Waiting for the task: (returnval){ [ 846.907023] env[61868]: value = "task-1315417" [ 846.907023] env[61868]: _type = "Task" [ 846.907023] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.913786] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 846.913786] env[61868]: value = "task-1315418" [ 846.913786] env[61868]: _type = "Task" [ 846.913786] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.919965] env[61868]: DEBUG oslo_vmware.api [None req-05a0085d-ae8b-4e73-9106-330b108967d5 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': task-1315417, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.925433] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315418, 'name': CreateVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.973764] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315412, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.072631] env[61868]: DEBUG nova.network.neutron [-] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.112507] env[61868]: DEBUG nova.compute.manager [req-00660052-504a-4f42-b5ac-4c974a6767d8 req-f1ea9a67-6b2a-4860-88ff-4a8c472f90e1 service nova] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Received event network-changed-29d18263-e917-46dc-8b91-33365c0d3481 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 847.112816] env[61868]: DEBUG nova.compute.manager [req-00660052-504a-4f42-b5ac-4c974a6767d8 req-f1ea9a67-6b2a-4860-88ff-4a8c472f90e1 service nova] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Refreshing instance network info cache due to event network-changed-29d18263-e917-46dc-8b91-33365c0d3481. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 847.113158] env[61868]: DEBUG oslo_concurrency.lockutils [req-00660052-504a-4f42-b5ac-4c974a6767d8 req-f1ea9a67-6b2a-4860-88ff-4a8c472f90e1 service nova] Acquiring lock "refresh_cache-409cad1a-946d-4c58-aa57-1c0bf97fe63f" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.113440] env[61868]: DEBUG oslo_concurrency.lockutils [req-00660052-504a-4f42-b5ac-4c974a6767d8 req-f1ea9a67-6b2a-4860-88ff-4a8c472f90e1 service nova] Acquired lock "refresh_cache-409cad1a-946d-4c58-aa57-1c0bf97fe63f" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.113676] env[61868]: DEBUG nova.network.neutron [req-00660052-504a-4f42-b5ac-4c974a6767d8 req-f1ea9a67-6b2a-4860-88ff-4a8c472f90e1 service nova] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Refreshing network info cache for port 29d18263-e917-46dc-8b91-33365c0d3481 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 847.117039] env[61868]: DEBUG oslo_concurrency.lockutils [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.190353] env[61868]: DEBUG oslo_vmware.api [None req-fd684283-1baf-4fb4-90ff-bd8128b2a028 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': task-1315413, 'name': PowerOffVM_Task, 'duration_secs': 0.32228} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.190687] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd684283-1baf-4fb4-90ff-bd8128b2a028 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 847.190936] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-fd684283-1baf-4fb4-90ff-bd8128b2a028 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 847.191370] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e97f336a-1557-4da6-b3f0-b1dfbf012705 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.232511] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315414, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067874} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.232794] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 847.233590] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c97e8d0-8ee5-46c3-bada-07a1ade5a33d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.257103] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] c2693a43-4ea2-4ab0-8915-2fa544780e3c/c2693a43-4ea2-4ab0-8915-2fa544780e3c.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 847.258169] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a2e7b77d-f02d-4fa5-a9c7-18a7c471a3e3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.277259] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-fd684283-1baf-4fb4-90ff-bd8128b2a028 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 847.277631] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-fd684283-1baf-4fb4-90ff-bd8128b2a028 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 847.277776] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd684283-1baf-4fb4-90ff-bd8128b2a028 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Deleting the datastore file [datastore1] 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 847.278062] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7f4e1e17-e2d3-40b8-a729-b628a4f97d71 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.281593] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Waiting for the task: (returnval){ [ 847.281593] env[61868]: value = "task-1315420" [ 847.281593] env[61868]: _type = "Task" [ 847.281593] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.286710] env[61868]: DEBUG oslo_vmware.api [None req-fd684283-1baf-4fb4-90ff-bd8128b2a028 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Waiting for the task: (returnval){ [ 847.286710] env[61868]: value = "task-1315421" [ 847.286710] env[61868]: _type = "Task" [ 847.286710] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.293723] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315420, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.300580] env[61868]: DEBUG oslo_vmware.api [None req-fd684283-1baf-4fb4-90ff-bd8128b2a028 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': task-1315421, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.322601] env[61868]: DEBUG nova.network.neutron [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Successfully updated port: 7b94f787-9d7e-4edf-92b0-44d37a7e93ec {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 847.335951] env[61868]: DEBUG nova.compute.utils [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 847.337174] env[61868]: DEBUG nova.compute.manager [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 847.337471] env[61868]: DEBUG nova.network.neutron [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 847.387876] env[61868]: DEBUG nova.policy [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '54d21acf9a57434697110d3c308ff708', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0204eb5697a943b7899aeb8e30c28e05', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 847.418219] env[61868]: DEBUG oslo_vmware.api [None req-05a0085d-ae8b-4e73-9106-330b108967d5 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': task-1315417, 'name': PowerOffVM_Task, 'duration_secs': 0.231887} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.423340] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-05a0085d-ae8b-4e73-9106-330b108967d5 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 847.423558] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-05a0085d-ae8b-4e73-9106-330b108967d5 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 847.423792] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6cbd41e7-12f6-48e7-a50c-7dab1c74cca2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.430743] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315418, 'name': CreateVM_Task, 'duration_secs': 0.355014} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.430930] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 847.431746] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.431934] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.432371] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 847.432618] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b8bf3da-cbe1-47be-b0ee-eb324a09d5f3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.437496] env[61868]: DEBUG oslo_vmware.api [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Waiting for the task: (returnval){ [ 847.437496] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52656523-b1a7-61ee-0d30-4f81f91fa673" [ 847.437496] env[61868]: _type = "Task" [ 847.437496] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.448284] env[61868]: DEBUG oslo_vmware.api [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52656523-b1a7-61ee-0d30-4f81f91fa673, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.474947] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315412, 'name': ReconfigVM_Task, 'duration_secs': 0.562171} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.475495] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Reconfigured VM instance instance-0000003c to attach disk [datastore2] 52a5afc6-4e79-436c-bc94-b61ca9fb860c/52a5afc6-4e79-436c-bc94-b61ca9fb860c.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 847.476566] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d93efa22-11aa-466f-92c4-7b23c77a3513 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.483758] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Waiting for the task: (returnval){ [ 847.483758] env[61868]: value = "task-1315423" [ 847.483758] env[61868]: _type = "Task" [ 847.483758] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.490940] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-05a0085d-ae8b-4e73-9106-330b108967d5 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 847.491400] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-05a0085d-ae8b-4e73-9106-330b108967d5 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Deleting contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 847.491728] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-05a0085d-ae8b-4e73-9106-330b108967d5 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Deleting the datastore file [datastore2] 7788f9f8-c08e-46f8-b204-070215223436 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 847.492180] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bb699bfd-a3f6-4975-8c85-809fc71c0d1e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.498980] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315423, 'name': Rename_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.504435] env[61868]: DEBUG oslo_vmware.api [None req-05a0085d-ae8b-4e73-9106-330b108967d5 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Waiting for the task: (returnval){ [ 847.504435] env[61868]: value = "task-1315424" [ 847.504435] env[61868]: _type = "Task" [ 847.504435] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.513430] env[61868]: DEBUG oslo_vmware.api [None req-05a0085d-ae8b-4e73-9106-330b108967d5 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': task-1315424, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.576240] env[61868]: INFO nova.compute.manager [-] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Took 1.03 seconds to deallocate network for instance. [ 847.710282] env[61868]: DEBUG nova.network.neutron [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Successfully created port: 9196ece9-3dc0-411e-b759-7e201fa02c56 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 847.798740] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315420, 'name': ReconfigVM_Task, 'duration_secs': 0.327532} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.799080] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Reconfigured VM instance instance-0000003e to attach disk [datastore2] c2693a43-4ea2-4ab0-8915-2fa544780e3c/c2693a43-4ea2-4ab0-8915-2fa544780e3c.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 847.803868] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cbd88d84-1829-4d53-8804-b2827dc3729e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.805608] env[61868]: DEBUG oslo_vmware.api [None req-fd684283-1baf-4fb4-90ff-bd8128b2a028 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': task-1315421, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.179843} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.805857] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd684283-1baf-4fb4-90ff-bd8128b2a028 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 847.806067] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-fd684283-1baf-4fb4-90ff-bd8128b2a028 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 847.806275] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-fd684283-1baf-4fb4-90ff-bd8128b2a028 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 847.806460] env[61868]: INFO nova.compute.manager [None req-fd684283-1baf-4fb4-90ff-bd8128b2a028 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Took 1.15 seconds to destroy the instance on the hypervisor. [ 847.806701] env[61868]: DEBUG oslo.service.loopingcall [None req-fd684283-1baf-4fb4-90ff-bd8128b2a028 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 847.807302] env[61868]: DEBUG nova.compute.manager [-] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 847.807421] env[61868]: DEBUG nova.network.neutron [-] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 847.811602] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Waiting for the task: (returnval){ [ 847.811602] env[61868]: value = "task-1315425" [ 847.811602] env[61868]: _type = "Task" [ 847.811602] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.820222] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315425, 'name': Rename_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.825881] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "refresh_cache-c7b10943-5136-44c8-b4e7-59651220a333" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.826044] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquired lock "refresh_cache-c7b10943-5136-44c8-b4e7-59651220a333" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.826194] env[61868]: DEBUG nova.network.neutron [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 847.843228] env[61868]: DEBUG nova.compute.manager [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 847.948735] env[61868]: DEBUG oslo_vmware.api [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52656523-b1a7-61ee-0d30-4f81f91fa673, 'name': SearchDatastore_Task, 'duration_secs': 0.010621} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.952184] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.952556] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 847.952879] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.953057] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.953253] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 847.953931] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b1c0c866-e878-4038-a2dd-5aed768e157b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.970445] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 847.970445] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 847.971758] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4cfd2b4-891d-49a0-80a8-168fc9d18b49 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.980267] env[61868]: DEBUG oslo_vmware.api [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Waiting for the task: (returnval){ [ 847.980267] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]526657f0-9648-4bd7-dfef-2fa5083fc9a3" [ 847.980267] env[61868]: _type = "Task" [ 847.980267] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.988422] env[61868]: DEBUG oslo_vmware.api [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]526657f0-9648-4bd7-dfef-2fa5083fc9a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.990098] env[61868]: DEBUG nova.network.neutron [req-00660052-504a-4f42-b5ac-4c974a6767d8 req-f1ea9a67-6b2a-4860-88ff-4a8c472f90e1 service nova] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Updated VIF entry in instance network info cache for port 29d18263-e917-46dc-8b91-33365c0d3481. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 847.990170] env[61868]: DEBUG nova.network.neutron [req-00660052-504a-4f42-b5ac-4c974a6767d8 req-f1ea9a67-6b2a-4860-88ff-4a8c472f90e1 service nova] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Updating instance_info_cache with network_info: [{"id": "29d18263-e917-46dc-8b91-33365c0d3481", "address": "fa:16:3e:c6:07:cd", "network": {"id": "0fa2f8e1-ee57-450f-9d06-9a056566585b", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-729551994-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7471600a510e4feeb0c819d0cca5eb48", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap29d18263-e9", "ovs_interfaceid": "29d18263-e917-46dc-8b91-33365c0d3481", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.003412] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315423, 'name': Rename_Task, 'duration_secs': 0.200138} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.004752] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 848.005124] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a7e95992-8287-4bfb-9524-959e370f72d0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.021752] env[61868]: DEBUG oslo_vmware.api [None req-05a0085d-ae8b-4e73-9106-330b108967d5 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Task: {'id': task-1315424, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165531} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.025754] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-05a0085d-ae8b-4e73-9106-330b108967d5 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 848.026058] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-05a0085d-ae8b-4e73-9106-330b108967d5 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Deleted contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 848.026299] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-05a0085d-ae8b-4e73-9106-330b108967d5 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 848.026485] env[61868]: INFO nova.compute.manager [None req-05a0085d-ae8b-4e73-9106-330b108967d5 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Took 1.17 seconds to destroy the instance on the hypervisor. [ 848.026791] env[61868]: DEBUG oslo.service.loopingcall [None req-05a0085d-ae8b-4e73-9106-330b108967d5 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 848.026992] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Waiting for the task: (returnval){ [ 848.026992] env[61868]: value = "task-1315426" [ 848.026992] env[61868]: _type = "Task" [ 848.026992] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.027449] env[61868]: DEBUG nova.compute.manager [-] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 848.027513] env[61868]: DEBUG nova.network.neutron [-] [instance: 7788f9f8-c08e-46f8-b204-070215223436] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 848.038140] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315426, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.088552] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.252867] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef4bc08e-156f-4df7-a8b8-4d3aecc69033 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.267066] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98e3e4d6-1f16-45d5-a901-8c00453851b5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.306215] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5503b83-ff96-42e6-8ee1-90ff69b52e09 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.310761] env[61868]: DEBUG nova.compute.manager [req-cc587891-7059-459a-90c2-fa3241c46d16 req-c85c2c73-1a05-44d4-9a3d-d8d2175c53fb service nova] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Received event network-vif-deleted-6b7fa3c1-e436-4fed-92f3-d40988d5c79c {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 848.310872] env[61868]: INFO nova.compute.manager [req-cc587891-7059-459a-90c2-fa3241c46d16 req-c85c2c73-1a05-44d4-9a3d-d8d2175c53fb service nova] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Neutron deleted interface 6b7fa3c1-e436-4fed-92f3-d40988d5c79c; detaching it from the instance and deleting it from the info cache [ 848.311080] env[61868]: DEBUG nova.network.neutron [req-cc587891-7059-459a-90c2-fa3241c46d16 req-c85c2c73-1a05-44d4-9a3d-d8d2175c53fb service nova] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.322210] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8edee686-da92-4f22-8419-6cd953505720 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.332050] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315425, 'name': Rename_Task, 'duration_secs': 0.148401} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.332953] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 848.333683] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b56f1db7-50aa-4095-b60a-2365f8e9b3bd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.344063] env[61868]: DEBUG nova.compute.provider_tree [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 848.351024] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Waiting for the task: (returnval){ [ 848.351024] env[61868]: value = "task-1315427" [ 848.351024] env[61868]: _type = "Task" [ 848.351024] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.363705] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315427, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.368993] env[61868]: DEBUG nova.network.neutron [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 848.489687] env[61868]: DEBUG oslo_vmware.api [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]526657f0-9648-4bd7-dfef-2fa5083fc9a3, 'name': SearchDatastore_Task, 'duration_secs': 0.008725} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.492764] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f29907c-54be-4183-837c-8322526c816b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.495504] env[61868]: DEBUG oslo_concurrency.lockutils [req-00660052-504a-4f42-b5ac-4c974a6767d8 req-f1ea9a67-6b2a-4860-88ff-4a8c472f90e1 service nova] Releasing lock "refresh_cache-409cad1a-946d-4c58-aa57-1c0bf97fe63f" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.499039] env[61868]: DEBUG oslo_vmware.api [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Waiting for the task: (returnval){ [ 848.499039] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]525d6c5e-26dc-37ed-c743-679d2e299333" [ 848.499039] env[61868]: _type = "Task" [ 848.499039] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.506865] env[61868]: DEBUG oslo_vmware.api [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]525d6c5e-26dc-37ed-c743-679d2e299333, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.535489] env[61868]: DEBUG nova.network.neutron [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Updating instance_info_cache with network_info: [{"id": "7b94f787-9d7e-4edf-92b0-44d37a7e93ec", "address": "fa:16:3e:63:08:8e", "network": {"id": "c1f5bf03-438f-4acc-93fd-159bed7d4069", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2061742146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a60c3c20950e4156b64c3b4c61b9f0f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24b91e3b-50f1-4a16-b929-942e6b31b2d8", "external-id": "nsx-vlan-transportzone-34", "segmentation_id": 34, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b94f787-9d", "ovs_interfaceid": "7b94f787-9d7e-4edf-92b0-44d37a7e93ec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.540238] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315426, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.771377] env[61868]: DEBUG nova.network.neutron [-] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.813424] env[61868]: DEBUG nova.network.neutron [-] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.815633] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a22070d1-4908-4e72-ac42-b8ca429dcecb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.825310] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9958d3bc-a33b-497b-a5b1-29a9cf62fb44 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.847479] env[61868]: DEBUG nova.scheduler.client.report [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 848.861909] env[61868]: DEBUG nova.compute.manager [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 848.865099] env[61868]: DEBUG nova.compute.manager [req-cc587891-7059-459a-90c2-fa3241c46d16 req-c85c2c73-1a05-44d4-9a3d-d8d2175c53fb service nova] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Detach interface failed, port_id=6b7fa3c1-e436-4fed-92f3-d40988d5c79c, reason: Instance 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc could not be found. {{(pid=61868) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 848.874112] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315427, 'name': PowerOnVM_Task, 'duration_secs': 0.458598} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.874375] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 848.874583] env[61868]: INFO nova.compute.manager [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Took 6.96 seconds to spawn the instance on the hypervisor. [ 848.874824] env[61868]: DEBUG nova.compute.manager [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 848.875635] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-157d0e48-2c9c-4058-a8b6-b66fa733cd6c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.892166] env[61868]: DEBUG nova.virt.hardware [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 848.892442] env[61868]: DEBUG nova.virt.hardware [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 848.892614] env[61868]: DEBUG nova.virt.hardware [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 848.892828] env[61868]: DEBUG nova.virt.hardware [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 848.893010] env[61868]: DEBUG nova.virt.hardware [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 848.893175] env[61868]: DEBUG nova.virt.hardware [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 848.893424] env[61868]: DEBUG nova.virt.hardware [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 848.893610] env[61868]: DEBUG nova.virt.hardware [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 848.893789] env[61868]: DEBUG nova.virt.hardware [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 848.893959] env[61868]: DEBUG nova.virt.hardware [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 848.894154] env[61868]: DEBUG nova.virt.hardware [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 848.894997] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93b35a3a-0948-42e2-916a-c4b672cd388e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.902888] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8fd3fce-0204-4bd3-9146-14ef515c37c8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.011580] env[61868]: DEBUG oslo_vmware.api [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]525d6c5e-26dc-37ed-c743-679d2e299333, 'name': SearchDatastore_Task, 'duration_secs': 0.012376} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.011580] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.012711] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 409cad1a-946d-4c58-aa57-1c0bf97fe63f/409cad1a-946d-4c58-aa57-1c0bf97fe63f.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 849.014225] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bbf932da-e53b-4c8a-880b-fe63fd93f407 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.020204] env[61868]: DEBUG oslo_vmware.api [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Waiting for the task: (returnval){ [ 849.020204] env[61868]: value = "task-1315428" [ 849.020204] env[61868]: _type = "Task" [ 849.020204] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.033306] env[61868]: DEBUG oslo_vmware.api [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315428, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.041731] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Releasing lock "refresh_cache-c7b10943-5136-44c8-b4e7-59651220a333" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.042066] env[61868]: DEBUG nova.compute.manager [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Instance network_info: |[{"id": "7b94f787-9d7e-4edf-92b0-44d37a7e93ec", "address": "fa:16:3e:63:08:8e", "network": {"id": "c1f5bf03-438f-4acc-93fd-159bed7d4069", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2061742146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a60c3c20950e4156b64c3b4c61b9f0f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24b91e3b-50f1-4a16-b929-942e6b31b2d8", "external-id": "nsx-vlan-transportzone-34", "segmentation_id": 34, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b94f787-9d", "ovs_interfaceid": "7b94f787-9d7e-4edf-92b0-44d37a7e93ec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 849.042403] env[61868]: DEBUG oslo_vmware.api [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315426, 'name': PowerOnVM_Task, 'duration_secs': 0.648878} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.042746] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:63:08:8e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24b91e3b-50f1-4a16-b929-942e6b31b2d8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7b94f787-9d7e-4edf-92b0-44d37a7e93ec', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 849.050095] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Creating folder: Project (a60c3c20950e4156b64c3b4c61b9f0f2). Parent ref: group-v281478. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 849.050426] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 849.050639] env[61868]: INFO nova.compute.manager [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Took 9.63 seconds to spawn the instance on the hypervisor. [ 849.050815] env[61868]: DEBUG nova.compute.manager [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 849.051420] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3d9476b3-7d5b-4fee-a23c-c222f3f21513 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.054033] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e05efad9-a440-44c0-93f8-ea6d23af4388 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.064978] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Created folder: Project (a60c3c20950e4156b64c3b4c61b9f0f2) in parent group-v281478. [ 849.064978] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Creating folder: Instances. Parent ref: group-v281523. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 849.065204] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2dfde93d-45d5-4ca3-8944-c92c33ab0e2e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.073502] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Created folder: Instances in parent group-v281523. [ 849.073767] env[61868]: DEBUG oslo.service.loopingcall [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 849.074527] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 849.074527] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-61fbceab-8eee-4fa2-adb6-48abfa5bf43e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.095532] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 849.095532] env[61868]: value = "task-1315431" [ 849.095532] env[61868]: _type = "Task" [ 849.095532] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.105517] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315431, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.156049] env[61868]: DEBUG nova.compute.manager [req-b3966aea-83eb-48a9-a0f9-f02fa52d30b4 req-075f22c0-11f7-47a5-a02f-b480fc97e2b7 service nova] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Received event network-vif-plugged-7b94f787-9d7e-4edf-92b0-44d37a7e93ec {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 849.156186] env[61868]: DEBUG oslo_concurrency.lockutils [req-b3966aea-83eb-48a9-a0f9-f02fa52d30b4 req-075f22c0-11f7-47a5-a02f-b480fc97e2b7 service nova] Acquiring lock "c7b10943-5136-44c8-b4e7-59651220a333-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.156411] env[61868]: DEBUG oslo_concurrency.lockutils [req-b3966aea-83eb-48a9-a0f9-f02fa52d30b4 req-075f22c0-11f7-47a5-a02f-b480fc97e2b7 service nova] Lock "c7b10943-5136-44c8-b4e7-59651220a333-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.156658] env[61868]: DEBUG oslo_concurrency.lockutils [req-b3966aea-83eb-48a9-a0f9-f02fa52d30b4 req-075f22c0-11f7-47a5-a02f-b480fc97e2b7 service nova] Lock "c7b10943-5136-44c8-b4e7-59651220a333-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.156794] env[61868]: DEBUG nova.compute.manager [req-b3966aea-83eb-48a9-a0f9-f02fa52d30b4 req-075f22c0-11f7-47a5-a02f-b480fc97e2b7 service nova] [instance: c7b10943-5136-44c8-b4e7-59651220a333] No waiting events found dispatching network-vif-plugged-7b94f787-9d7e-4edf-92b0-44d37a7e93ec {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 849.157068] env[61868]: WARNING nova.compute.manager [req-b3966aea-83eb-48a9-a0f9-f02fa52d30b4 req-075f22c0-11f7-47a5-a02f-b480fc97e2b7 service nova] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Received unexpected event network-vif-plugged-7b94f787-9d7e-4edf-92b0-44d37a7e93ec for instance with vm_state building and task_state spawning. [ 849.157191] env[61868]: DEBUG nova.compute.manager [req-b3966aea-83eb-48a9-a0f9-f02fa52d30b4 req-075f22c0-11f7-47a5-a02f-b480fc97e2b7 service nova] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Received event network-changed-7b94f787-9d7e-4edf-92b0-44d37a7e93ec {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 849.157332] env[61868]: DEBUG nova.compute.manager [req-b3966aea-83eb-48a9-a0f9-f02fa52d30b4 req-075f22c0-11f7-47a5-a02f-b480fc97e2b7 service nova] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Refreshing instance network info cache due to event network-changed-7b94f787-9d7e-4edf-92b0-44d37a7e93ec. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 849.157529] env[61868]: DEBUG oslo_concurrency.lockutils [req-b3966aea-83eb-48a9-a0f9-f02fa52d30b4 req-075f22c0-11f7-47a5-a02f-b480fc97e2b7 service nova] Acquiring lock "refresh_cache-c7b10943-5136-44c8-b4e7-59651220a333" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.157743] env[61868]: DEBUG oslo_concurrency.lockutils [req-b3966aea-83eb-48a9-a0f9-f02fa52d30b4 req-075f22c0-11f7-47a5-a02f-b480fc97e2b7 service nova] Acquired lock "refresh_cache-c7b10943-5136-44c8-b4e7-59651220a333" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.157894] env[61868]: DEBUG nova.network.neutron [req-b3966aea-83eb-48a9-a0f9-f02fa52d30b4 req-075f22c0-11f7-47a5-a02f-b480fc97e2b7 service nova] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Refreshing network info cache for port 7b94f787-9d7e-4edf-92b0-44d37a7e93ec {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 849.274253] env[61868]: INFO nova.compute.manager [-] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Took 1.47 seconds to deallocate network for instance. [ 849.316127] env[61868]: INFO nova.compute.manager [-] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Took 1.29 seconds to deallocate network for instance. [ 849.352305] env[61868]: DEBUG oslo_concurrency.lockutils [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.520s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.352889] env[61868]: DEBUG nova.compute.manager [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 849.356138] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 6.335s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.356236] env[61868]: DEBUG nova.objects.instance [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61868) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 849.395217] env[61868]: DEBUG nova.network.neutron [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Successfully updated port: 9196ece9-3dc0-411e-b759-7e201fa02c56 {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 849.397571] env[61868]: INFO nova.compute.manager [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Took 24.19 seconds to build instance. [ 849.530112] env[61868]: DEBUG oslo_vmware.api [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315428, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.574928] env[61868]: INFO nova.compute.manager [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Took 25.48 seconds to build instance. [ 849.607274] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315431, 'name': CreateVM_Task, 'duration_secs': 0.412359} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.608049] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 849.608206] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.608450] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.608790] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 849.609086] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-21e951db-453f-4686-805b-13e083cd61a9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.615250] env[61868]: DEBUG oslo_vmware.api [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 849.615250] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52126b8e-ebc7-2a97-6c6b-99f56f4261e9" [ 849.615250] env[61868]: _type = "Task" [ 849.615250] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.624135] env[61868]: DEBUG oslo_vmware.api [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52126b8e-ebc7-2a97-6c6b-99f56f4261e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.783172] env[61868]: DEBUG oslo_concurrency.lockutils [None req-fd684283-1baf-4fb4-90ff-bd8128b2a028 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.823638] env[61868]: DEBUG oslo_concurrency.lockutils [None req-05a0085d-ae8b-4e73-9106-330b108967d5 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.860838] env[61868]: DEBUG nova.compute.utils [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 849.865461] env[61868]: DEBUG nova.compute.manager [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 849.865461] env[61868]: DEBUG nova.network.neutron [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 849.896798] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Acquiring lock "refresh_cache-ca85dfd1-e794-41dc-a337-fbd99998e1fa" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.897066] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Acquired lock "refresh_cache-ca85dfd1-e794-41dc-a337-fbd99998e1fa" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.897331] env[61868]: DEBUG nova.network.neutron [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 849.899539] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Lock "c2693a43-4ea2-4ab0-8915-2fa544780e3c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 114.729s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.909021] env[61868]: DEBUG nova.network.neutron [req-b3966aea-83eb-48a9-a0f9-f02fa52d30b4 req-075f22c0-11f7-47a5-a02f-b480fc97e2b7 service nova] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Updated VIF entry in instance network info cache for port 7b94f787-9d7e-4edf-92b0-44d37a7e93ec. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 849.909021] env[61868]: DEBUG nova.network.neutron [req-b3966aea-83eb-48a9-a0f9-f02fa52d30b4 req-075f22c0-11f7-47a5-a02f-b480fc97e2b7 service nova] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Updating instance_info_cache with network_info: [{"id": "7b94f787-9d7e-4edf-92b0-44d37a7e93ec", "address": "fa:16:3e:63:08:8e", "network": {"id": "c1f5bf03-438f-4acc-93fd-159bed7d4069", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2061742146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a60c3c20950e4156b64c3b4c61b9f0f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24b91e3b-50f1-4a16-b929-942e6b31b2d8", "external-id": "nsx-vlan-transportzone-34", "segmentation_id": 34, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b94f787-9d", "ovs_interfaceid": "7b94f787-9d7e-4edf-92b0-44d37a7e93ec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.928666] env[61868]: DEBUG nova.policy [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1425850b70c940988169260ea59c21df', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7541757d43c74b93acf728aa2fb0f425', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 850.032555] env[61868]: DEBUG oslo_vmware.api [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315428, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.079351] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c4272cd0-f569-40be-86e4-89f3720809b2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Lock "52a5afc6-4e79-436c-bc94-b61ca9fb860c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 114.939s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.125313] env[61868]: DEBUG oslo_vmware.api [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52126b8e-ebc7-2a97-6c6b-99f56f4261e9, 'name': SearchDatastore_Task, 'duration_secs': 0.014575} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.125706] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.125955] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 850.126284] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.126441] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.126745] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 850.127085] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-122d5c70-0027-4e15-8cf8-104853a72b1e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.136751] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 850.137065] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 850.137766] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-677d1c0b-4c5e-48b9-8403-36d700afc385 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.143322] env[61868]: DEBUG oslo_vmware.api [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 850.143322] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]525be2bc-403a-1e6c-566e-8cf8e812479a" [ 850.143322] env[61868]: _type = "Task" [ 850.143322] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.153899] env[61868]: DEBUG oslo_vmware.api [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]525be2bc-403a-1e6c-566e-8cf8e812479a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.202374] env[61868]: DEBUG nova.network.neutron [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Successfully created port: 77081d78-be71-47a0-a7be-8511ae0243d4 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 850.368490] env[61868]: DEBUG nova.compute.manager [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 850.370138] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ce40a022-b28b-46a2-b260-f1167c3146c3 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.371689] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.842s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.374124] env[61868]: INFO nova.compute.claims [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 850.405550] env[61868]: DEBUG nova.compute.manager [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 850.412678] env[61868]: DEBUG oslo_concurrency.lockutils [req-b3966aea-83eb-48a9-a0f9-f02fa52d30b4 req-075f22c0-11f7-47a5-a02f-b480fc97e2b7 service nova] Releasing lock "refresh_cache-c7b10943-5136-44c8-b4e7-59651220a333" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.416019] env[61868]: DEBUG nova.compute.manager [req-b3966aea-83eb-48a9-a0f9-f02fa52d30b4 req-075f22c0-11f7-47a5-a02f-b480fc97e2b7 service nova] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Received event network-vif-deleted-29d0d831-eb07-403f-bde9-16844b987447 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 850.440602] env[61868]: DEBUG oslo_concurrency.lockutils [None req-241ac167-bf6d-4de7-a078-a1e9812e3546 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Acquiring lock "3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.440602] env[61868]: DEBUG oslo_concurrency.lockutils [None req-241ac167-bf6d-4de7-a078-a1e9812e3546 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Lock "3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.440602] env[61868]: DEBUG oslo_concurrency.lockutils [None req-241ac167-bf6d-4de7-a078-a1e9812e3546 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Acquiring lock "3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.440602] env[61868]: DEBUG oslo_concurrency.lockutils [None req-241ac167-bf6d-4de7-a078-a1e9812e3546 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Lock "3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.441142] env[61868]: DEBUG oslo_concurrency.lockutils [None req-241ac167-bf6d-4de7-a078-a1e9812e3546 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Lock "3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.442272] env[61868]: INFO nova.compute.manager [None req-241ac167-bf6d-4de7-a078-a1e9812e3546 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Terminating instance [ 850.445016] env[61868]: DEBUG nova.compute.manager [None req-241ac167-bf6d-4de7-a078-a1e9812e3546 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 850.445444] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-241ac167-bf6d-4de7-a078-a1e9812e3546 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 850.446619] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a397f71-ab7b-4a11-8cb7-8c1a43a4a734 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.451111] env[61868]: DEBUG nova.network.neutron [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 850.456876] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-241ac167-bf6d-4de7-a078-a1e9812e3546 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 850.457299] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2cf82aa4-79f0-4361-8b49-585d524592c5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.464074] env[61868]: DEBUG oslo_vmware.api [None req-241ac167-bf6d-4de7-a078-a1e9812e3546 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Waiting for the task: (returnval){ [ 850.464074] env[61868]: value = "task-1315432" [ 850.464074] env[61868]: _type = "Task" [ 850.464074] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.474415] env[61868]: DEBUG oslo_vmware.api [None req-241ac167-bf6d-4de7-a078-a1e9812e3546 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315432, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.535909] env[61868]: DEBUG oslo_vmware.api [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315428, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.582702] env[61868]: DEBUG nova.compute.manager [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 850.633551] env[61868]: DEBUG nova.network.neutron [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Updating instance_info_cache with network_info: [{"id": "9196ece9-3dc0-411e-b759-7e201fa02c56", "address": "fa:16:3e:d9:83:54", "network": {"id": "c70832d7-da02-4109-a493-f55ec3d083ca", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1165780467-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0204eb5697a943b7899aeb8e30c28e05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f0ef5aba-bd9a-42ff-a1a0-5e763986d70a", "external-id": "nsx-vlan-transportzone-209", "segmentation_id": 209, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9196ece9-3d", "ovs_interfaceid": "9196ece9-3dc0-411e-b759-7e201fa02c56", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.661866] env[61868]: DEBUG oslo_vmware.api [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]525be2bc-403a-1e6c-566e-8cf8e812479a, 'name': SearchDatastore_Task, 'duration_secs': 0.015459} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.663190] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee3d30c8-252c-4b89-be16-b8d40fdcbdc0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.672681] env[61868]: DEBUG oslo_vmware.api [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 850.672681] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]524028b8-fa03-b5f2-6135-60097ebcaf01" [ 850.672681] env[61868]: _type = "Task" [ 850.672681] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.686840] env[61868]: DEBUG oslo_vmware.api [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]524028b8-fa03-b5f2-6135-60097ebcaf01, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.929127] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.974644] env[61868]: DEBUG oslo_vmware.api [None req-241ac167-bf6d-4de7-a078-a1e9812e3546 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315432, 'name': PowerOffVM_Task, 'duration_secs': 0.228373} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.974932] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-241ac167-bf6d-4de7-a078-a1e9812e3546 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 850.975114] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-241ac167-bf6d-4de7-a078-a1e9812e3546 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 850.975367] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d39e0e6e-60df-4e6a-be72-4a6632745605 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.033470] env[61868]: DEBUG oslo_vmware.api [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315428, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.808321} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.033732] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 409cad1a-946d-4c58-aa57-1c0bf97fe63f/409cad1a-946d-4c58-aa57-1c0bf97fe63f.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 851.033961] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 851.034272] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cc563e68-a4cd-468a-b1d0-3979e9aff102 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.041119] env[61868]: DEBUG oslo_vmware.api [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Waiting for the task: (returnval){ [ 851.041119] env[61868]: value = "task-1315434" [ 851.041119] env[61868]: _type = "Task" [ 851.041119] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.045899] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-241ac167-bf6d-4de7-a078-a1e9812e3546 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 851.046131] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-241ac167-bf6d-4de7-a078-a1e9812e3546 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Deleting contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 851.046347] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-241ac167-bf6d-4de7-a078-a1e9812e3546 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Deleting the datastore file [datastore2] 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 851.047015] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6fba52a7-8cec-46fe-80c7-dfee434f9cac {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.051672] env[61868]: DEBUG oslo_vmware.api [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315434, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.056367] env[61868]: DEBUG oslo_vmware.api [None req-241ac167-bf6d-4de7-a078-a1e9812e3546 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Waiting for the task: (returnval){ [ 851.056367] env[61868]: value = "task-1315435" [ 851.056367] env[61868]: _type = "Task" [ 851.056367] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.064360] env[61868]: DEBUG oslo_vmware.api [None req-241ac167-bf6d-4de7-a078-a1e9812e3546 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315435, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.113969] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.136810] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Releasing lock "refresh_cache-ca85dfd1-e794-41dc-a337-fbd99998e1fa" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.137195] env[61868]: DEBUG nova.compute.manager [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Instance network_info: |[{"id": "9196ece9-3dc0-411e-b759-7e201fa02c56", "address": "fa:16:3e:d9:83:54", "network": {"id": "c70832d7-da02-4109-a493-f55ec3d083ca", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1165780467-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0204eb5697a943b7899aeb8e30c28e05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f0ef5aba-bd9a-42ff-a1a0-5e763986d70a", "external-id": "nsx-vlan-transportzone-209", "segmentation_id": 209, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9196ece9-3d", "ovs_interfaceid": "9196ece9-3dc0-411e-b759-7e201fa02c56", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 851.137670] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d9:83:54', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f0ef5aba-bd9a-42ff-a1a0-5e763986d70a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9196ece9-3dc0-411e-b759-7e201fa02c56', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 851.145466] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Creating folder: Project (0204eb5697a943b7899aeb8e30c28e05). Parent ref: group-v281478. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 851.145788] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cd32af2c-131b-44ec-80c0-7f6a08fcc1dd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.157585] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Created folder: Project (0204eb5697a943b7899aeb8e30c28e05) in parent group-v281478. [ 851.157829] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Creating folder: Instances. Parent ref: group-v281526. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 851.158104] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9a590fa5-945c-4a6a-bee3-1d69d1331752 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.168258] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Created folder: Instances in parent group-v281526. [ 851.169054] env[61868]: DEBUG oslo.service.loopingcall [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 851.169054] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 851.169054] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-47e277c5-e56c-4a71-93b1-f935ef51676b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.197333] env[61868]: DEBUG oslo_vmware.api [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]524028b8-fa03-b5f2-6135-60097ebcaf01, 'name': SearchDatastore_Task, 'duration_secs': 0.024807} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.198557] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.198940] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] c7b10943-5136-44c8-b4e7-59651220a333/c7b10943-5136-44c8-b4e7-59651220a333.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 851.199244] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 851.199244] env[61868]: value = "task-1315438" [ 851.199244] env[61868]: _type = "Task" [ 851.199244] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.199533] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ad4659b5-0e3f-45b4-8e68-3855a1ba9480 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.210796] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315438, 'name': CreateVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.212672] env[61868]: DEBUG oslo_vmware.api [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 851.212672] env[61868]: value = "task-1315439" [ 851.212672] env[61868]: _type = "Task" [ 851.212672] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.221714] env[61868]: DEBUG oslo_vmware.api [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315439, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.268501] env[61868]: DEBUG nova.compute.manager [req-8987c75d-10a5-4f6d-8c52-9f81811abef4 req-ec41c396-f869-44a8-8f78-43e7a587c6ff service nova] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Received event network-vif-plugged-9196ece9-3dc0-411e-b759-7e201fa02c56 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 851.268939] env[61868]: DEBUG oslo_concurrency.lockutils [req-8987c75d-10a5-4f6d-8c52-9f81811abef4 req-ec41c396-f869-44a8-8f78-43e7a587c6ff service nova] Acquiring lock "ca85dfd1-e794-41dc-a337-fbd99998e1fa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.269333] env[61868]: DEBUG oslo_concurrency.lockutils [req-8987c75d-10a5-4f6d-8c52-9f81811abef4 req-ec41c396-f869-44a8-8f78-43e7a587c6ff service nova] Lock "ca85dfd1-e794-41dc-a337-fbd99998e1fa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.269684] env[61868]: DEBUG oslo_concurrency.lockutils [req-8987c75d-10a5-4f6d-8c52-9f81811abef4 req-ec41c396-f869-44a8-8f78-43e7a587c6ff service nova] Lock "ca85dfd1-e794-41dc-a337-fbd99998e1fa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.269997] env[61868]: DEBUG nova.compute.manager [req-8987c75d-10a5-4f6d-8c52-9f81811abef4 req-ec41c396-f869-44a8-8f78-43e7a587c6ff service nova] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] No waiting events found dispatching network-vif-plugged-9196ece9-3dc0-411e-b759-7e201fa02c56 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 851.270313] env[61868]: WARNING nova.compute.manager [req-8987c75d-10a5-4f6d-8c52-9f81811abef4 req-ec41c396-f869-44a8-8f78-43e7a587c6ff service nova] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Received unexpected event network-vif-plugged-9196ece9-3dc0-411e-b759-7e201fa02c56 for instance with vm_state building and task_state spawning. [ 851.270643] env[61868]: DEBUG nova.compute.manager [req-8987c75d-10a5-4f6d-8c52-9f81811abef4 req-ec41c396-f869-44a8-8f78-43e7a587c6ff service nova] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Received event network-changed-9196ece9-3dc0-411e-b759-7e201fa02c56 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 851.270914] env[61868]: DEBUG nova.compute.manager [req-8987c75d-10a5-4f6d-8c52-9f81811abef4 req-ec41c396-f869-44a8-8f78-43e7a587c6ff service nova] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Refreshing instance network info cache due to event network-changed-9196ece9-3dc0-411e-b759-7e201fa02c56. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 851.271201] env[61868]: DEBUG oslo_concurrency.lockutils [req-8987c75d-10a5-4f6d-8c52-9f81811abef4 req-ec41c396-f869-44a8-8f78-43e7a587c6ff service nova] Acquiring lock "refresh_cache-ca85dfd1-e794-41dc-a337-fbd99998e1fa" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.271389] env[61868]: DEBUG oslo_concurrency.lockutils [req-8987c75d-10a5-4f6d-8c52-9f81811abef4 req-ec41c396-f869-44a8-8f78-43e7a587c6ff service nova] Acquired lock "refresh_cache-ca85dfd1-e794-41dc-a337-fbd99998e1fa" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.271584] env[61868]: DEBUG nova.network.neutron [req-8987c75d-10a5-4f6d-8c52-9f81811abef4 req-ec41c396-f869-44a8-8f78-43e7a587c6ff service nova] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Refreshing network info cache for port 9196ece9-3dc0-411e-b759-7e201fa02c56 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 851.378799] env[61868]: DEBUG nova.compute.manager [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 851.412711] env[61868]: DEBUG nova.virt.hardware [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 851.413154] env[61868]: DEBUG nova.virt.hardware [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 851.413221] env[61868]: DEBUG nova.virt.hardware [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 851.413545] env[61868]: DEBUG nova.virt.hardware [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 851.413616] env[61868]: DEBUG nova.virt.hardware [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 851.413811] env[61868]: DEBUG nova.virt.hardware [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 851.414296] env[61868]: DEBUG nova.virt.hardware [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 851.414485] env[61868]: DEBUG nova.virt.hardware [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 851.415714] env[61868]: DEBUG nova.virt.hardware [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 851.415931] env[61868]: DEBUG nova.virt.hardware [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 851.416890] env[61868]: DEBUG nova.virt.hardware [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 851.417472] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-298ad6dd-e338-4d81-b52d-c25de8ffc5da {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.431407] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9d8255a-7dd7-4c3d-b9c0-2ff724cc5a44 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.553474] env[61868]: DEBUG oslo_vmware.api [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315434, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086624} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.556527] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 851.558048] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c90ebad0-662d-45d7-8d65-a1229a4eee00 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.570948] env[61868]: DEBUG oslo_vmware.api [None req-241ac167-bf6d-4de7-a078-a1e9812e3546 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315435, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.295493} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.584165] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-241ac167-bf6d-4de7-a078-a1e9812e3546 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 851.584165] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-241ac167-bf6d-4de7-a078-a1e9812e3546 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Deleted contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 851.584407] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-241ac167-bf6d-4de7-a078-a1e9812e3546 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 851.584494] env[61868]: INFO nova.compute.manager [None req-241ac167-bf6d-4de7-a078-a1e9812e3546 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Took 1.14 seconds to destroy the instance on the hypervisor. [ 851.584795] env[61868]: DEBUG oslo.service.loopingcall [None req-241ac167-bf6d-4de7-a078-a1e9812e3546 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 851.594411] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Reconfiguring VM instance instance-0000003f to attach disk [datastore1] 409cad1a-946d-4c58-aa57-1c0bf97fe63f/409cad1a-946d-4c58-aa57-1c0bf97fe63f.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 851.595158] env[61868]: DEBUG nova.compute.manager [-] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 851.595158] env[61868]: DEBUG nova.network.neutron [-] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 851.597170] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8171854d-fbff-4ae6-b9b4-31e3f88ab6e9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.620710] env[61868]: DEBUG oslo_vmware.api [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Waiting for the task: (returnval){ [ 851.620710] env[61868]: value = "task-1315440" [ 851.620710] env[61868]: _type = "Task" [ 851.620710] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.631915] env[61868]: DEBUG oslo_vmware.api [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315440, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.722593] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315438, 'name': CreateVM_Task, 'duration_secs': 0.462557} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.722895] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 851.724340] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.724592] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.724985] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 851.727530] env[61868]: DEBUG oslo_vmware.rw_handles [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5276e24e-54b4-9464-2978-4c5f0e74aec0/disk-0.vmdk. {{(pid=61868) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 851.730682] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de07a76f-f4f4-4749-b00c-51c0b5f02ade {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.733307] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8446f29-af68-450d-b6a3-c9f0fb3dcd9f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.741718] env[61868]: DEBUG oslo_vmware.api [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315439, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.745271] env[61868]: DEBUG oslo_vmware.api [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Waiting for the task: (returnval){ [ 851.745271] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]527eb2d0-072c-16d4-6dcb-8a2dd84ad7ca" [ 851.745271] env[61868]: _type = "Task" [ 851.745271] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.747232] env[61868]: DEBUG oslo_vmware.rw_handles [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5276e24e-54b4-9464-2978-4c5f0e74aec0/disk-0.vmdk is in state: ready. {{(pid=61868) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 851.747484] env[61868]: ERROR oslo_vmware.rw_handles [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5276e24e-54b4-9464-2978-4c5f0e74aec0/disk-0.vmdk due to incomplete transfer. [ 851.751048] env[61868]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-9bfffb2a-8988-442e-8b1e-103a293855d2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.760684] env[61868]: DEBUG oslo_vmware.api [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]527eb2d0-072c-16d4-6dcb-8a2dd84ad7ca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.762352] env[61868]: DEBUG oslo_vmware.rw_handles [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5276e24e-54b4-9464-2978-4c5f0e74aec0/disk-0.vmdk. {{(pid=61868) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 851.762637] env[61868]: DEBUG nova.virt.vmwareapi.images [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Uploaded image 577e8014-2c7d-4375-ad66-8f0ffb82b743 to the Glance image server {{(pid=61868) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 851.765039] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Destroying the VM {{(pid=61868) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 851.765945] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-82dd284d-98c3-4a98-8e01-6f7f62156970 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.772713] env[61868]: DEBUG oslo_vmware.api [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 851.772713] env[61868]: value = "task-1315441" [ 851.772713] env[61868]: _type = "Task" [ 851.772713] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.788972] env[61868]: DEBUG oslo_vmware.api [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315441, 'name': Destroy_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.840079] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a55e8fa9-e2ca-40ab-b03c-40f04fe8f1c7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.848247] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b30590de-613e-4ca1-83b0-82eddc2c94fa {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.880782] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3d40a55-da0b-4a15-8c6d-aeb4fd12d9db {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.888775] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-505e512d-ff43-4563-bad0-2ac83ac378ca {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.903416] env[61868]: DEBUG nova.compute.provider_tree [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 851.993098] env[61868]: DEBUG nova.network.neutron [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Successfully updated port: 77081d78-be71-47a0-a7be-8511ae0243d4 {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 852.130348] env[61868]: DEBUG oslo_vmware.api [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315440, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.226104] env[61868]: DEBUG oslo_vmware.api [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315439, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.587788} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.228964] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] c7b10943-5136-44c8-b4e7-59651220a333/c7b10943-5136-44c8-b4e7-59651220a333.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 852.228964] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 852.228964] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-840f0ecf-6889-4357-b725-82f797a07a76 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.233490] env[61868]: DEBUG oslo_vmware.api [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 852.233490] env[61868]: value = "task-1315442" [ 852.233490] env[61868]: _type = "Task" [ 852.233490] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.241511] env[61868]: DEBUG oslo_vmware.api [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315442, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.256297] env[61868]: DEBUG oslo_vmware.api [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]527eb2d0-072c-16d4-6dcb-8a2dd84ad7ca, 'name': SearchDatastore_Task, 'duration_secs': 0.021199} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.256609] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.256890] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 852.257150] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.257308] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.257505] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 852.257775] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b9a3a971-cbc9-4c74-a868-40bd436ef8c6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.266826] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 852.267035] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 852.267758] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-95dc5ea4-5212-442a-b211-b2ce8d9005e1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.273475] env[61868]: DEBUG oslo_vmware.api [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Waiting for the task: (returnval){ [ 852.273475] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52d3ecc8-7d26-7e50-ef3a-29e2b3ece595" [ 852.273475] env[61868]: _type = "Task" [ 852.273475] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.293128] env[61868]: DEBUG oslo_vmware.api [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315441, 'name': Destroy_Task, 'duration_secs': 0.449625} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.297555] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Destroyed the VM [ 852.297996] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Deleting Snapshot of the VM instance {{(pid=61868) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 852.298684] env[61868]: DEBUG oslo_vmware.api [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52d3ecc8-7d26-7e50-ef3a-29e2b3ece595, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.298939] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-7266fc87-24da-4b6b-8fdf-0a5943245b34 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.306838] env[61868]: DEBUG oslo_vmware.api [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 852.306838] env[61868]: value = "task-1315443" [ 852.306838] env[61868]: _type = "Task" [ 852.306838] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.314585] env[61868]: DEBUG oslo_vmware.api [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315443, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.379434] env[61868]: DEBUG nova.network.neutron [req-8987c75d-10a5-4f6d-8c52-9f81811abef4 req-ec41c396-f869-44a8-8f78-43e7a587c6ff service nova] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Updated VIF entry in instance network info cache for port 9196ece9-3dc0-411e-b759-7e201fa02c56. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 852.379906] env[61868]: DEBUG nova.network.neutron [req-8987c75d-10a5-4f6d-8c52-9f81811abef4 req-ec41c396-f869-44a8-8f78-43e7a587c6ff service nova] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Updating instance_info_cache with network_info: [{"id": "9196ece9-3dc0-411e-b759-7e201fa02c56", "address": "fa:16:3e:d9:83:54", "network": {"id": "c70832d7-da02-4109-a493-f55ec3d083ca", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1165780467-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0204eb5697a943b7899aeb8e30c28e05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f0ef5aba-bd9a-42ff-a1a0-5e763986d70a", "external-id": "nsx-vlan-transportzone-209", "segmentation_id": 209, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9196ece9-3d", "ovs_interfaceid": "9196ece9-3dc0-411e-b759-7e201fa02c56", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.406835] env[61868]: DEBUG nova.scheduler.client.report [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 852.495816] env[61868]: DEBUG oslo_concurrency.lockutils [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Acquiring lock "refresh_cache-19c2720c-90bc-47f6-999b-6031f893408d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.495984] env[61868]: DEBUG oslo_concurrency.lockutils [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Acquired lock "refresh_cache-19c2720c-90bc-47f6-999b-6031f893408d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.496152] env[61868]: DEBUG nova.network.neutron [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 852.526769] env[61868]: DEBUG nova.network.neutron [-] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.631846] env[61868]: DEBUG oslo_vmware.api [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315440, 'name': ReconfigVM_Task, 'duration_secs': 0.601138} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.631846] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Reconfigured VM instance instance-0000003f to attach disk [datastore1] 409cad1a-946d-4c58-aa57-1c0bf97fe63f/409cad1a-946d-4c58-aa57-1c0bf97fe63f.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 852.632044] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-42629d4e-1732-44e3-9e6b-c4f072cb818f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.642531] env[61868]: DEBUG oslo_vmware.api [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Waiting for the task: (returnval){ [ 852.642531] env[61868]: value = "task-1315444" [ 852.642531] env[61868]: _type = "Task" [ 852.642531] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.650532] env[61868]: DEBUG oslo_vmware.api [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315444, 'name': Rename_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.743107] env[61868]: DEBUG oslo_vmware.api [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315442, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.147202} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.743393] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 852.744149] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7910f8a4-c265-44c0-bb42-c4e0b4c4b787 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.769013] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] c7b10943-5136-44c8-b4e7-59651220a333/c7b10943-5136-44c8-b4e7-59651220a333.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 852.769317] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e9b84291-b1b1-4862-ae4e-ac382d60f316 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.791952] env[61868]: DEBUG oslo_vmware.api [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52d3ecc8-7d26-7e50-ef3a-29e2b3ece595, 'name': SearchDatastore_Task, 'duration_secs': 0.018812} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.793652] env[61868]: DEBUG oslo_vmware.api [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 852.793652] env[61868]: value = "task-1315445" [ 852.793652] env[61868]: _type = "Task" [ 852.793652] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.793857] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-617ebcca-3988-41fd-b1c4-ffae7c1558ac {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.802062] env[61868]: DEBUG oslo_vmware.api [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Waiting for the task: (returnval){ [ 852.802062] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52465a21-35f5-7429-979a-00944e2f75f2" [ 852.802062] env[61868]: _type = "Task" [ 852.802062] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.805028] env[61868]: DEBUG oslo_vmware.api [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315445, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.815154] env[61868]: DEBUG oslo_vmware.api [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315443, 'name': RemoveSnapshot_Task, 'duration_secs': 0.347767} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.818058] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Deleted Snapshot of the VM instance {{(pid=61868) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 852.818347] env[61868]: DEBUG nova.compute.manager [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 852.818621] env[61868]: DEBUG oslo_vmware.api [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52465a21-35f5-7429-979a-00944e2f75f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.819311] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ea6bb98-8ebe-41c5-8f2b-b92468c1dd0a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.883143] env[61868]: DEBUG oslo_concurrency.lockutils [req-8987c75d-10a5-4f6d-8c52-9f81811abef4 req-ec41c396-f869-44a8-8f78-43e7a587c6ff service nova] Releasing lock "refresh_cache-ca85dfd1-e794-41dc-a337-fbd99998e1fa" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.911649] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.540s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.912163] env[61868]: DEBUG nova.compute.manager [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 852.914740] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.326s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.916618] env[61868]: INFO nova.compute.claims [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 853.030624] env[61868]: INFO nova.compute.manager [-] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Took 1.44 seconds to deallocate network for instance. [ 853.033630] env[61868]: DEBUG nova.network.neutron [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 853.154324] env[61868]: DEBUG oslo_vmware.api [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315444, 'name': Rename_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.173287] env[61868]: DEBUG nova.network.neutron [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Updating instance_info_cache with network_info: [{"id": "77081d78-be71-47a0-a7be-8511ae0243d4", "address": "fa:16:3e:79:ff:17", "network": {"id": "094eaf15-c932-4d4f-8736-fe5239160507", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1905020504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "7541757d43c74b93acf728aa2fb0f425", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap77081d78-be", "ovs_interfaceid": "77081d78-be71-47a0-a7be-8511ae0243d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.294714] env[61868]: DEBUG nova.compute.manager [req-ced25e4b-3c42-4673-835f-3ee191b46263 req-4e75bcd1-68d8-4277-93cb-b825da958ad8 service nova] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Received event network-vif-plugged-77081d78-be71-47a0-a7be-8511ae0243d4 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 853.294991] env[61868]: DEBUG oslo_concurrency.lockutils [req-ced25e4b-3c42-4673-835f-3ee191b46263 req-4e75bcd1-68d8-4277-93cb-b825da958ad8 service nova] Acquiring lock "19c2720c-90bc-47f6-999b-6031f893408d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.295280] env[61868]: DEBUG oslo_concurrency.lockutils [req-ced25e4b-3c42-4673-835f-3ee191b46263 req-4e75bcd1-68d8-4277-93cb-b825da958ad8 service nova] Lock "19c2720c-90bc-47f6-999b-6031f893408d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.295503] env[61868]: DEBUG oslo_concurrency.lockutils [req-ced25e4b-3c42-4673-835f-3ee191b46263 req-4e75bcd1-68d8-4277-93cb-b825da958ad8 service nova] Lock "19c2720c-90bc-47f6-999b-6031f893408d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.295727] env[61868]: DEBUG nova.compute.manager [req-ced25e4b-3c42-4673-835f-3ee191b46263 req-4e75bcd1-68d8-4277-93cb-b825da958ad8 service nova] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] No waiting events found dispatching network-vif-plugged-77081d78-be71-47a0-a7be-8511ae0243d4 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 853.295952] env[61868]: WARNING nova.compute.manager [req-ced25e4b-3c42-4673-835f-3ee191b46263 req-4e75bcd1-68d8-4277-93cb-b825da958ad8 service nova] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Received unexpected event network-vif-plugged-77081d78-be71-47a0-a7be-8511ae0243d4 for instance with vm_state building and task_state spawning. [ 853.296383] env[61868]: DEBUG nova.compute.manager [req-ced25e4b-3c42-4673-835f-3ee191b46263 req-4e75bcd1-68d8-4277-93cb-b825da958ad8 service nova] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Received event network-changed-77081d78-be71-47a0-a7be-8511ae0243d4 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 853.296383] env[61868]: DEBUG nova.compute.manager [req-ced25e4b-3c42-4673-835f-3ee191b46263 req-4e75bcd1-68d8-4277-93cb-b825da958ad8 service nova] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Refreshing instance network info cache due to event network-changed-77081d78-be71-47a0-a7be-8511ae0243d4. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 853.296610] env[61868]: DEBUG oslo_concurrency.lockutils [req-ced25e4b-3c42-4673-835f-3ee191b46263 req-4e75bcd1-68d8-4277-93cb-b825da958ad8 service nova] Acquiring lock "refresh_cache-19c2720c-90bc-47f6-999b-6031f893408d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.308415] env[61868]: DEBUG oslo_vmware.api [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315445, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.316895] env[61868]: DEBUG oslo_vmware.api [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52465a21-35f5-7429-979a-00944e2f75f2, 'name': SearchDatastore_Task, 'duration_secs': 0.050057} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.317127] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.317428] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] ca85dfd1-e794-41dc-a337-fbd99998e1fa/ca85dfd1-e794-41dc-a337-fbd99998e1fa.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 853.317744] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-72f2a4c9-e3a5-4ee9-b766-dbe91e68ecf5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.325400] env[61868]: DEBUG oslo_vmware.api [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Waiting for the task: (returnval){ [ 853.325400] env[61868]: value = "task-1315446" [ 853.325400] env[61868]: _type = "Task" [ 853.325400] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.336456] env[61868]: DEBUG oslo_vmware.api [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Task: {'id': task-1315446, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.336925] env[61868]: INFO nova.compute.manager [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Shelve offloading [ 853.338645] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 853.338879] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ab8864e7-6a75-4a2e-bc66-6dd01e860517 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.344983] env[61868]: DEBUG oslo_vmware.api [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 853.344983] env[61868]: value = "task-1315447" [ 853.344983] env[61868]: _type = "Task" [ 853.344983] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.353131] env[61868]: DEBUG oslo_vmware.api [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315447, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.421352] env[61868]: DEBUG nova.compute.utils [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 853.424757] env[61868]: DEBUG nova.compute.manager [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 853.424935] env[61868]: DEBUG nova.network.neutron [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 853.483850] env[61868]: DEBUG nova.policy [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3c44b4da602b4022b643fc49d72f2d22', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2b9b800086764b41a8817d1d8e463a2f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 853.543265] env[61868]: DEBUG oslo_concurrency.lockutils [None req-241ac167-bf6d-4de7-a078-a1e9812e3546 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.653817] env[61868]: DEBUG oslo_vmware.api [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315444, 'name': Rename_Task, 'duration_secs': 0.587783} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.654183] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 853.654458] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8919a57c-b50f-4d87-a56c-f78ab56873c9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.661032] env[61868]: DEBUG oslo_vmware.api [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Waiting for the task: (returnval){ [ 853.661032] env[61868]: value = "task-1315448" [ 853.661032] env[61868]: _type = "Task" [ 853.661032] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.668701] env[61868]: DEBUG oslo_vmware.api [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315448, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.676337] env[61868]: DEBUG oslo_concurrency.lockutils [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Releasing lock "refresh_cache-19c2720c-90bc-47f6-999b-6031f893408d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.676660] env[61868]: DEBUG nova.compute.manager [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Instance network_info: |[{"id": "77081d78-be71-47a0-a7be-8511ae0243d4", "address": "fa:16:3e:79:ff:17", "network": {"id": "094eaf15-c932-4d4f-8736-fe5239160507", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1905020504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "7541757d43c74b93acf728aa2fb0f425", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap77081d78-be", "ovs_interfaceid": "77081d78-be71-47a0-a7be-8511ae0243d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 853.676960] env[61868]: DEBUG oslo_concurrency.lockutils [req-ced25e4b-3c42-4673-835f-3ee191b46263 req-4e75bcd1-68d8-4277-93cb-b825da958ad8 service nova] Acquired lock "refresh_cache-19c2720c-90bc-47f6-999b-6031f893408d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.678422] env[61868]: DEBUG nova.network.neutron [req-ced25e4b-3c42-4673-835f-3ee191b46263 req-4e75bcd1-68d8-4277-93cb-b825da958ad8 service nova] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Refreshing network info cache for port 77081d78-be71-47a0-a7be-8511ae0243d4 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 853.680207] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:79:ff:17', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69f65356-c85e-4b7f-ad28-7c7b5e8cf50c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '77081d78-be71-47a0-a7be-8511ae0243d4', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 853.687887] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Creating folder: Project (7541757d43c74b93acf728aa2fb0f425). Parent ref: group-v281478. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 853.690919] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8c134792-79fd-4ed6-a811-03de63111804 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.702483] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Created folder: Project (7541757d43c74b93acf728aa2fb0f425) in parent group-v281478. [ 853.703512] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Creating folder: Instances. Parent ref: group-v281529. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 853.704799] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b7f83b5f-7c0e-42b0-a4dd-cb8a04dec004 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.715901] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Created folder: Instances in parent group-v281529. [ 853.716224] env[61868]: DEBUG oslo.service.loopingcall [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 853.716475] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 853.716733] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3a5039a6-7f1d-45f8-a5aa-c48ddbe1f96a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.738927] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 853.738927] env[61868]: value = "task-1315451" [ 853.738927] env[61868]: _type = "Task" [ 853.738927] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.747085] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315451, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.808235] env[61868]: DEBUG oslo_vmware.api [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315445, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.836332] env[61868]: DEBUG oslo_vmware.api [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Task: {'id': task-1315446, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.854664] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] VM already powered off {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 853.854890] env[61868]: DEBUG nova.compute.manager [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 853.855687] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56347149-bd68-41bc-a095-7a5478bbbc1e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.861589] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "refresh_cache-a16a726b-9673-4de7-9188-19c608a7dc0b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.861773] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquired lock "refresh_cache-a16a726b-9673-4de7-9188-19c608a7dc0b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.861953] env[61868]: DEBUG nova.network.neutron [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 853.926167] env[61868]: DEBUG nova.compute.manager [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 853.929423] env[61868]: DEBUG nova.network.neutron [req-ced25e4b-3c42-4673-835f-3ee191b46263 req-4e75bcd1-68d8-4277-93cb-b825da958ad8 service nova] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Updated VIF entry in instance network info cache for port 77081d78-be71-47a0-a7be-8511ae0243d4. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 853.929823] env[61868]: DEBUG nova.network.neutron [req-ced25e4b-3c42-4673-835f-3ee191b46263 req-4e75bcd1-68d8-4277-93cb-b825da958ad8 service nova] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Updating instance_info_cache with network_info: [{"id": "77081d78-be71-47a0-a7be-8511ae0243d4", "address": "fa:16:3e:79:ff:17", "network": {"id": "094eaf15-c932-4d4f-8736-fe5239160507", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1905020504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "7541757d43c74b93acf728aa2fb0f425", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap77081d78-be", "ovs_interfaceid": "77081d78-be71-47a0-a7be-8511ae0243d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.973492] env[61868]: DEBUG nova.network.neutron [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Successfully created port: 21d6cab8-f691-425c-bfa8-9d2f1c294b64 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 854.171280] env[61868]: DEBUG oslo_vmware.api [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315448, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.250468] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315451, 'name': CreateVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.252929] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79401205-4a2c-4979-a237-2582c3a1ea0d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.260697] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa8608b0-5395-4464-8f86-0e6d56be07c9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.293772] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-258dfbb6-367c-4175-9017-baa598c6c9d9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.311469] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-920eb1be-9c21-4f45-aef1-9ebbfb89b0b3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.315558] env[61868]: DEBUG oslo_vmware.api [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315445, 'name': ReconfigVM_Task, 'duration_secs': 1.074675} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.316140] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Reconfigured VM instance instance-00000040 to attach disk [datastore2] c7b10943-5136-44c8-b4e7-59651220a333/c7b10943-5136-44c8-b4e7-59651220a333.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 854.317270] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ef9bdf23-534b-456f-8991-17e1786926c1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.327737] env[61868]: DEBUG nova.compute.provider_tree [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 854.334345] env[61868]: DEBUG oslo_vmware.api [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 854.334345] env[61868]: value = "task-1315452" [ 854.334345] env[61868]: _type = "Task" [ 854.334345] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.345174] env[61868]: DEBUG oslo_vmware.api [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Task: {'id': task-1315446, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.351235] env[61868]: DEBUG oslo_vmware.api [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315452, 'name': Rename_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.437618] env[61868]: DEBUG oslo_concurrency.lockutils [req-ced25e4b-3c42-4673-835f-3ee191b46263 req-4e75bcd1-68d8-4277-93cb-b825da958ad8 service nova] Releasing lock "refresh_cache-19c2720c-90bc-47f6-999b-6031f893408d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.437618] env[61868]: DEBUG nova.compute.manager [req-ced25e4b-3c42-4673-835f-3ee191b46263 req-4e75bcd1-68d8-4277-93cb-b825da958ad8 service nova] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Received event network-vif-deleted-ec4c8961-6ef8-4b0e-bb38-99c843e9fd4c {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 854.594817] env[61868]: DEBUG nova.network.neutron [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Updating instance_info_cache with network_info: [{"id": "548fe1c2-07a5-45a2-b084-cc95216cdcad", "address": "fa:16:3e:99:ec:d5", "network": {"id": "5b4c5906-b873-4369-855b-e0898ad84812", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-254512637-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "812e115cc1114d0e8536924c4f5556f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap548fe1c2-07", "ovs_interfaceid": "548fe1c2-07a5-45a2-b084-cc95216cdcad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.673282] env[61868]: DEBUG oslo_vmware.api [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315448, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.751023] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315451, 'name': CreateVM_Task, 'duration_secs': 0.571139} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.751122] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 854.751836] env[61868]: DEBUG oslo_concurrency.lockutils [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.752048] env[61868]: DEBUG oslo_concurrency.lockutils [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.752381] env[61868]: DEBUG oslo_concurrency.lockutils [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 854.752667] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-feda22aa-f4d2-45f5-815f-e463bdcf99d1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.757464] env[61868]: DEBUG oslo_vmware.api [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 854.757464] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52a96246-8c74-5abb-4b46-6a29944ef24a" [ 854.757464] env[61868]: _type = "Task" [ 854.757464] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.764976] env[61868]: DEBUG oslo_vmware.api [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52a96246-8c74-5abb-4b46-6a29944ef24a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.832225] env[61868]: DEBUG nova.scheduler.client.report [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 854.843933] env[61868]: DEBUG oslo_vmware.api [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Task: {'id': task-1315446, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.370519} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.844657] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] ca85dfd1-e794-41dc-a337-fbd99998e1fa/ca85dfd1-e794-41dc-a337-fbd99998e1fa.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 854.844824] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 854.845104] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1ff4970f-38ba-4a63-9269-3df1e0207349 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.849617] env[61868]: DEBUG oslo_vmware.api [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315452, 'name': Rename_Task, 'duration_secs': 0.159127} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.850767] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 854.851015] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-92ec7db5-38a8-4d5c-ac4f-4307d8189d07 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.854451] env[61868]: DEBUG oslo_vmware.api [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Waiting for the task: (returnval){ [ 854.854451] env[61868]: value = "task-1315453" [ 854.854451] env[61868]: _type = "Task" [ 854.854451] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.858479] env[61868]: DEBUG oslo_vmware.api [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 854.858479] env[61868]: value = "task-1315454" [ 854.858479] env[61868]: _type = "Task" [ 854.858479] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.865986] env[61868]: DEBUG oslo_vmware.api [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315454, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.868932] env[61868]: DEBUG oslo_vmware.api [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Task: {'id': task-1315453, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.943978] env[61868]: DEBUG nova.compute.manager [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 854.969876] env[61868]: DEBUG nova.virt.hardware [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 854.969876] env[61868]: DEBUG nova.virt.hardware [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 854.969876] env[61868]: DEBUG nova.virt.hardware [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 854.970160] env[61868]: DEBUG nova.virt.hardware [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 854.970160] env[61868]: DEBUG nova.virt.hardware [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 854.970160] env[61868]: DEBUG nova.virt.hardware [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 854.970285] env[61868]: DEBUG nova.virt.hardware [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 854.970443] env[61868]: DEBUG nova.virt.hardware [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 854.970863] env[61868]: DEBUG nova.virt.hardware [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 854.970863] env[61868]: DEBUG nova.virt.hardware [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 854.971164] env[61868]: DEBUG nova.virt.hardware [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 854.971845] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d80ea85a-450e-454d-b6cf-0dc088e46e92 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.980161] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8b3463b-2008-47e0-b06e-aaf1cff6a0bc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.100809] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Releasing lock "refresh_cache-a16a726b-9673-4de7-9188-19c608a7dc0b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.173794] env[61868]: DEBUG oslo_vmware.api [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315448, 'name': PowerOnVM_Task, 'duration_secs': 1.46738} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.174088] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 855.174298] env[61868]: INFO nova.compute.manager [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Took 10.85 seconds to spawn the instance on the hypervisor. [ 855.174478] env[61868]: DEBUG nova.compute.manager [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 855.175352] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aebe0e0-ef9e-4ac1-ba36-e7c91d48779f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.273666] env[61868]: DEBUG oslo_vmware.api [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52a96246-8c74-5abb-4b46-6a29944ef24a, 'name': SearchDatastore_Task, 'duration_secs': 0.020832} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.273666] env[61868]: DEBUG oslo_concurrency.lockutils [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.273666] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 855.273666] env[61868]: DEBUG oslo_concurrency.lockutils [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.273820] env[61868]: DEBUG oslo_concurrency.lockutils [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.273820] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 855.275087] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b8deece9-cdff-4d90-b0c3-254b572ee7d1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.286226] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 855.286489] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 855.287689] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-48c71577-9ec1-44ef-8ee1-a1850f54c43e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.297436] env[61868]: DEBUG oslo_vmware.api [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 855.297436] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5289bf3e-8a60-fbd8-56ef-5e3d45ce7868" [ 855.297436] env[61868]: _type = "Task" [ 855.297436] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.310263] env[61868]: DEBUG oslo_vmware.api [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5289bf3e-8a60-fbd8-56ef-5e3d45ce7868, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.338445] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.424s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.339071] env[61868]: DEBUG nova.compute.manager [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 855.341873] env[61868]: DEBUG oslo_concurrency.lockutils [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.225s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.343283] env[61868]: INFO nova.compute.claims [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 855.380013] env[61868]: DEBUG oslo_vmware.api [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315454, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.380325] env[61868]: DEBUG oslo_vmware.api [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Task: {'id': task-1315453, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063772} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.380581] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 855.382011] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d3aa7bf-d113-42c6-9965-8e1fd8e731f8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.387065] env[61868]: DEBUG nova.compute.manager [req-1b193c35-eba2-4640-a1b0-57c8c434302f req-c80c825a-ddf8-4af0-ab93-ab32e941d22f service nova] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Received event network-vif-unplugged-548fe1c2-07a5-45a2-b084-cc95216cdcad {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 855.387274] env[61868]: DEBUG oslo_concurrency.lockutils [req-1b193c35-eba2-4640-a1b0-57c8c434302f req-c80c825a-ddf8-4af0-ab93-ab32e941d22f service nova] Acquiring lock "a16a726b-9673-4de7-9188-19c608a7dc0b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.387491] env[61868]: DEBUG oslo_concurrency.lockutils [req-1b193c35-eba2-4640-a1b0-57c8c434302f req-c80c825a-ddf8-4af0-ab93-ab32e941d22f service nova] Lock "a16a726b-9673-4de7-9188-19c608a7dc0b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.387664] env[61868]: DEBUG oslo_concurrency.lockutils [req-1b193c35-eba2-4640-a1b0-57c8c434302f req-c80c825a-ddf8-4af0-ab93-ab32e941d22f service nova] Lock "a16a726b-9673-4de7-9188-19c608a7dc0b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.387839] env[61868]: DEBUG nova.compute.manager [req-1b193c35-eba2-4640-a1b0-57c8c434302f req-c80c825a-ddf8-4af0-ab93-ab32e941d22f service nova] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] No waiting events found dispatching network-vif-unplugged-548fe1c2-07a5-45a2-b084-cc95216cdcad {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 855.388023] env[61868]: WARNING nova.compute.manager [req-1b193c35-eba2-4640-a1b0-57c8c434302f req-c80c825a-ddf8-4af0-ab93-ab32e941d22f service nova] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Received unexpected event network-vif-unplugged-548fe1c2-07a5-45a2-b084-cc95216cdcad for instance with vm_state shelved and task_state shelving_offloading. [ 855.409664] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] ca85dfd1-e794-41dc-a337-fbd99998e1fa/ca85dfd1-e794-41dc-a337-fbd99998e1fa.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 855.411786] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 855.412327] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ab438a7e-975c-4a8a-8976-3e54b1504237 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.427374] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1730b16d-ed0d-4ab1-923b-cb610ccab858 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.436182] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 855.437289] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9dcbe85b-96a3-4c02-b569-7e77ab685afc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.438997] env[61868]: DEBUG oslo_vmware.api [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Waiting for the task: (returnval){ [ 855.438997] env[61868]: value = "task-1315455" [ 855.438997] env[61868]: _type = "Task" [ 855.438997] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.447367] env[61868]: DEBUG oslo_vmware.api [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Task: {'id': task-1315455, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.601473] env[61868]: DEBUG nova.compute.manager [req-538c0bfb-42a7-4ed1-a657-832dd8844a6d req-23c81501-7de2-4ae5-8405-aacdc19abca9 service nova] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Received event network-vif-plugged-21d6cab8-f691-425c-bfa8-9d2f1c294b64 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 855.601810] env[61868]: DEBUG oslo_concurrency.lockutils [req-538c0bfb-42a7-4ed1-a657-832dd8844a6d req-23c81501-7de2-4ae5-8405-aacdc19abca9 service nova] Acquiring lock "75fdbc11-1ef3-442d-bcf4-55069ba9ab62-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.602107] env[61868]: DEBUG oslo_concurrency.lockutils [req-538c0bfb-42a7-4ed1-a657-832dd8844a6d req-23c81501-7de2-4ae5-8405-aacdc19abca9 service nova] Lock "75fdbc11-1ef3-442d-bcf4-55069ba9ab62-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.602304] env[61868]: DEBUG oslo_concurrency.lockutils [req-538c0bfb-42a7-4ed1-a657-832dd8844a6d req-23c81501-7de2-4ae5-8405-aacdc19abca9 service nova] Lock "75fdbc11-1ef3-442d-bcf4-55069ba9ab62-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.602486] env[61868]: DEBUG nova.compute.manager [req-538c0bfb-42a7-4ed1-a657-832dd8844a6d req-23c81501-7de2-4ae5-8405-aacdc19abca9 service nova] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] No waiting events found dispatching network-vif-plugged-21d6cab8-f691-425c-bfa8-9d2f1c294b64 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 855.602660] env[61868]: WARNING nova.compute.manager [req-538c0bfb-42a7-4ed1-a657-832dd8844a6d req-23c81501-7de2-4ae5-8405-aacdc19abca9 service nova] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Received unexpected event network-vif-plugged-21d6cab8-f691-425c-bfa8-9d2f1c294b64 for instance with vm_state building and task_state spawning. [ 855.693348] env[61868]: INFO nova.compute.manager [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Took 26.89 seconds to build instance. [ 855.810027] env[61868]: DEBUG oslo_vmware.api [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5289bf3e-8a60-fbd8-56ef-5e3d45ce7868, 'name': SearchDatastore_Task, 'duration_secs': 0.011987} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.810027] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6c7bc17-bf0f-4de7-a478-e88913854828 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.814519] env[61868]: DEBUG oslo_vmware.api [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 855.814519] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5299c1c0-8842-7ba8-8e48-ce818e9c7f01" [ 855.814519] env[61868]: _type = "Task" [ 855.814519] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.822414] env[61868]: DEBUG oslo_vmware.api [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5299c1c0-8842-7ba8-8e48-ce818e9c7f01, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.850026] env[61868]: DEBUG nova.compute.utils [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 855.850862] env[61868]: DEBUG nova.compute.manager [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 855.851184] env[61868]: DEBUG nova.network.neutron [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 855.872328] env[61868]: DEBUG oslo_vmware.api [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315454, 'name': PowerOnVM_Task, 'duration_secs': 0.949244} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.872328] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 855.872659] env[61868]: INFO nova.compute.manager [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Took 9.06 seconds to spawn the instance on the hypervisor. [ 855.872965] env[61868]: DEBUG nova.compute.manager [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 855.873929] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f23782c-3aa7-48a0-b973-1046b109cd50 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.891557] env[61868]: DEBUG nova.policy [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd6f85890c5824e3eaf1a0bd2b2f83eac', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '87aa9ce723574a31a0632d3736039ca4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 855.918688] env[61868]: DEBUG nova.network.neutron [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Successfully updated port: 21d6cab8-f691-425c-bfa8-9d2f1c294b64 {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 855.950040] env[61868]: DEBUG oslo_vmware.api [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Task: {'id': task-1315455, 'name': ReconfigVM_Task, 'duration_secs': 0.32914} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.950040] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Reconfigured VM instance instance-00000041 to attach disk [datastore1] ca85dfd1-e794-41dc-a337-fbd99998e1fa/ca85dfd1-e794-41dc-a337-fbd99998e1fa.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 855.952024] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f8697bc8-efbb-461e-b9ea-567ef9984753 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.957268] env[61868]: DEBUG oslo_vmware.api [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Waiting for the task: (returnval){ [ 855.957268] env[61868]: value = "task-1315457" [ 855.957268] env[61868]: _type = "Task" [ 855.957268] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.967506] env[61868]: DEBUG oslo_vmware.api [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Task: {'id': task-1315457, 'name': Rename_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.174029] env[61868]: DEBUG nova.network.neutron [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Successfully created port: 5159fed6-c58a-4dc0-9b7d-964aa28faaef {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 856.195925] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a22d0b9d-ebed-418a-b79a-dafdfc78c9ce tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Lock "409cad1a-946d-4c58-aa57-1c0bf97fe63f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.947s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.327852] env[61868]: DEBUG oslo_vmware.api [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5299c1c0-8842-7ba8-8e48-ce818e9c7f01, 'name': SearchDatastore_Task, 'duration_secs': 0.070532} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.328248] env[61868]: DEBUG oslo_concurrency.lockutils [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.328521] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 19c2720c-90bc-47f6-999b-6031f893408d/19c2720c-90bc-47f6-999b-6031f893408d.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 856.328797] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0139c635-1709-45a7-af88-1d55a134c05e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.336183] env[61868]: DEBUG oslo_vmware.api [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 856.336183] env[61868]: value = "task-1315458" [ 856.336183] env[61868]: _type = "Task" [ 856.336183] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.343350] env[61868]: DEBUG oslo_vmware.api [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315458, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.358084] env[61868]: DEBUG nova.compute.manager [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 856.396052] env[61868]: INFO nova.compute.manager [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Took 23.88 seconds to build instance. [ 856.424453] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquiring lock "refresh_cache-75fdbc11-1ef3-442d-bcf4-55069ba9ab62" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.424649] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquired lock "refresh_cache-75fdbc11-1ef3-442d-bcf4-55069ba9ab62" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.425041] env[61868]: DEBUG nova.network.neutron [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 856.467965] env[61868]: DEBUG oslo_vmware.api [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Task: {'id': task-1315457, 'name': Rename_Task, 'duration_secs': 0.138012} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.468360] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 856.468633] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-75775732-8187-460b-9465-8eac6f29d8af {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.478146] env[61868]: DEBUG oslo_vmware.api [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Waiting for the task: (returnval){ [ 856.478146] env[61868]: value = "task-1315459" [ 856.478146] env[61868]: _type = "Task" [ 856.478146] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.486519] env[61868]: DEBUG oslo_vmware.api [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Task: {'id': task-1315459, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.688551] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0728b10-9bd7-4ee6-a982-62b126947b05 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.697017] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ada436b-a49e-4a05-9da5-28333818f06d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.700679] env[61868]: DEBUG nova.compute.manager [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 856.735528] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d61b1d4-d0f4-4799-baad-0d7cb310f802 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.746022] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1b70926-1929-4e9d-8c96-72edfa75c68a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.763493] env[61868]: DEBUG nova.compute.provider_tree [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 856.847471] env[61868]: DEBUG oslo_vmware.api [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315458, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.903420] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9dadab01-5966-4503-9a65-15d9dc418569 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "c7b10943-5136-44c8-b4e7-59651220a333" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 111.863s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.995555] env[61868]: DEBUG oslo_vmware.api [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Task: {'id': task-1315459, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.996835] env[61868]: DEBUG nova.network.neutron [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 857.226030] env[61868]: DEBUG oslo_concurrency.lockutils [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.256548] env[61868]: DEBUG nova.network.neutron [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Updating instance_info_cache with network_info: [{"id": "21d6cab8-f691-425c-bfa8-9d2f1c294b64", "address": "fa:16:3e:76:a4:70", "network": {"id": "e7457f1a-171a-4c82-a260-1d1a7a95861e", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1032778554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b9b800086764b41a8817d1d8e463a2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21d6cab8-f6", "ovs_interfaceid": "21d6cab8-f691-425c-bfa8-9d2f1c294b64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.267718] env[61868]: DEBUG nova.scheduler.client.report [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 857.345955] env[61868]: DEBUG oslo_vmware.api [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315458, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.708388} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.346373] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 19c2720c-90bc-47f6-999b-6031f893408d/19c2720c-90bc-47f6-999b-6031f893408d.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 857.346606] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 857.346878] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-14fad861-c49b-43d7-8265-05e2df882279 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.353488] env[61868]: DEBUG oslo_vmware.api [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 857.353488] env[61868]: value = "task-1315460" [ 857.353488] env[61868]: _type = "Task" [ 857.353488] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.365808] env[61868]: DEBUG oslo_vmware.api [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315460, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.369311] env[61868]: DEBUG nova.compute.manager [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 857.406222] env[61868]: DEBUG nova.virt.hardware [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 857.406495] env[61868]: DEBUG nova.virt.hardware [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 857.406672] env[61868]: DEBUG nova.virt.hardware [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 857.406895] env[61868]: DEBUG nova.virt.hardware [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 857.408910] env[61868]: DEBUG nova.virt.hardware [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 857.409159] env[61868]: DEBUG nova.virt.hardware [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 857.409400] env[61868]: DEBUG nova.virt.hardware [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 857.409614] env[61868]: DEBUG nova.virt.hardware [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 857.410599] env[61868]: DEBUG nova.virt.hardware [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 857.410818] env[61868]: DEBUG nova.virt.hardware [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 857.411033] env[61868]: DEBUG nova.virt.hardware [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 857.411482] env[61868]: DEBUG nova.compute.manager [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 857.414718] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d2ceefa-7313-4433-b40e-49ec8036d3da {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.431632] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4498930b-5795-43b9-8b95-02a16f1b49ed {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.488296] env[61868]: DEBUG oslo_vmware.api [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Task: {'id': task-1315459, 'name': PowerOnVM_Task, 'duration_secs': 0.839694} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.488635] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 857.488771] env[61868]: INFO nova.compute.manager [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Took 8.63 seconds to spawn the instance on the hypervisor. [ 857.488955] env[61868]: DEBUG nova.compute.manager [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 857.489764] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66c910de-2ca0-4b3c-9227-9f0d6e487341 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.686164] env[61868]: DEBUG nova.compute.manager [req-5d944fa5-b95a-4ed5-a353-cec0232f5330 req-3d9f70f1-ca37-46c4-a372-be1c1fde6772 service nova] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Received event network-changed-548fe1c2-07a5-45a2-b084-cc95216cdcad {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 857.686370] env[61868]: DEBUG nova.compute.manager [req-5d944fa5-b95a-4ed5-a353-cec0232f5330 req-3d9f70f1-ca37-46c4-a372-be1c1fde6772 service nova] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Refreshing instance network info cache due to event network-changed-548fe1c2-07a5-45a2-b084-cc95216cdcad. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 857.686666] env[61868]: DEBUG oslo_concurrency.lockutils [req-5d944fa5-b95a-4ed5-a353-cec0232f5330 req-3d9f70f1-ca37-46c4-a372-be1c1fde6772 service nova] Acquiring lock "refresh_cache-a16a726b-9673-4de7-9188-19c608a7dc0b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.686743] env[61868]: DEBUG oslo_concurrency.lockutils [req-5d944fa5-b95a-4ed5-a353-cec0232f5330 req-3d9f70f1-ca37-46c4-a372-be1c1fde6772 service nova] Acquired lock "refresh_cache-a16a726b-9673-4de7-9188-19c608a7dc0b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.686970] env[61868]: DEBUG nova.network.neutron [req-5d944fa5-b95a-4ed5-a353-cec0232f5330 req-3d9f70f1-ca37-46c4-a372-be1c1fde6772 service nova] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Refreshing network info cache for port 548fe1c2-07a5-45a2-b084-cc95216cdcad {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 857.726535] env[61868]: DEBUG nova.network.neutron [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Successfully updated port: 5159fed6-c58a-4dc0-9b7d-964aa28faaef {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 857.759757] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Releasing lock "refresh_cache-75fdbc11-1ef3-442d-bcf4-55069ba9ab62" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.760611] env[61868]: DEBUG nova.compute.manager [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Instance network_info: |[{"id": "21d6cab8-f691-425c-bfa8-9d2f1c294b64", "address": "fa:16:3e:76:a4:70", "network": {"id": "e7457f1a-171a-4c82-a260-1d1a7a95861e", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1032778554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b9b800086764b41a8817d1d8e463a2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21d6cab8-f6", "ovs_interfaceid": "21d6cab8-f691-425c-bfa8-9d2f1c294b64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 857.760828] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:76:a4:70', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd19577c9-1b2e-490b-8031-2f278dd3f570', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '21d6cab8-f691-425c-bfa8-9d2f1c294b64', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 857.770478] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Creating folder: Project (2b9b800086764b41a8817d1d8e463a2f). Parent ref: group-v281478. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 857.770478] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-361934f9-4a46-464b-b9d4-9392f5893172 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.774388] env[61868]: DEBUG oslo_concurrency.lockutils [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.433s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.774881] env[61868]: DEBUG nova.compute.manager [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 857.780583] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.689s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.780583] env[61868]: DEBUG nova.objects.instance [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Lazy-loading 'resources' on Instance uuid 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 857.788845] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Created folder: Project (2b9b800086764b41a8817d1d8e463a2f) in parent group-v281478. [ 857.789060] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Creating folder: Instances. Parent ref: group-v281532. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 857.789302] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8b553bba-2533-4649-970b-000e13f7fe75 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.799838] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Created folder: Instances in parent group-v281532. [ 857.800112] env[61868]: DEBUG oslo.service.loopingcall [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 857.800310] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 857.800519] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c1e20784-a248-44be-9c78-e17e1204f854 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.817685] env[61868]: DEBUG nova.compute.manager [req-b6029332-0df1-40d1-be13-2bab39e19071 req-c35b1a84-fbd0-40a6-9986-712ce70a15f0 service nova] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Received event network-changed-21d6cab8-f691-425c-bfa8-9d2f1c294b64 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 857.817887] env[61868]: DEBUG nova.compute.manager [req-b6029332-0df1-40d1-be13-2bab39e19071 req-c35b1a84-fbd0-40a6-9986-712ce70a15f0 service nova] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Refreshing instance network info cache due to event network-changed-21d6cab8-f691-425c-bfa8-9d2f1c294b64. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 857.818117] env[61868]: DEBUG oslo_concurrency.lockutils [req-b6029332-0df1-40d1-be13-2bab39e19071 req-c35b1a84-fbd0-40a6-9986-712ce70a15f0 service nova] Acquiring lock "refresh_cache-75fdbc11-1ef3-442d-bcf4-55069ba9ab62" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.818266] env[61868]: DEBUG oslo_concurrency.lockutils [req-b6029332-0df1-40d1-be13-2bab39e19071 req-c35b1a84-fbd0-40a6-9986-712ce70a15f0 service nova] Acquired lock "refresh_cache-75fdbc11-1ef3-442d-bcf4-55069ba9ab62" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.818502] env[61868]: DEBUG nova.network.neutron [req-b6029332-0df1-40d1-be13-2bab39e19071 req-c35b1a84-fbd0-40a6-9986-712ce70a15f0 service nova] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Refreshing network info cache for port 21d6cab8-f691-425c-bfa8-9d2f1c294b64 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 857.824780] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 857.824780] env[61868]: value = "task-1315463" [ 857.824780] env[61868]: _type = "Task" [ 857.824780] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.833763] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315463, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.863703] env[61868]: DEBUG oslo_vmware.api [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315460, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06731} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.864430] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 857.865266] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d94f28a-c1ee-4e55-8e72-e02bdabbee0c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.889215] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] 19c2720c-90bc-47f6-999b-6031f893408d/19c2720c-90bc-47f6-999b-6031f893408d.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 857.889964] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fb429f9b-66e4-4ad9-9e56-eb7ee8000c2a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.911381] env[61868]: DEBUG oslo_vmware.api [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 857.911381] env[61868]: value = "task-1315464" [ 857.911381] env[61868]: _type = "Task" [ 857.911381] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.924091] env[61868]: DEBUG oslo_vmware.api [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315464, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.942929] env[61868]: DEBUG oslo_concurrency.lockutils [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.012594] env[61868]: INFO nova.compute.manager [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Took 24.74 seconds to build instance. [ 858.228030] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Acquiring lock "refresh_cache-c6f99cbc-3e81-44b4-b3d1-d8cd821829f2" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.228369] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Acquired lock "refresh_cache-c6f99cbc-3e81-44b4-b3d1-d8cd821829f2" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.228546] env[61868]: DEBUG nova.network.neutron [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 858.283076] env[61868]: DEBUG nova.compute.utils [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 858.287326] env[61868]: DEBUG nova.compute.manager [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Not allocating networking since 'none' was specified. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 858.335342] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315463, 'name': CreateVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.379401] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 858.379689] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 858.379897] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Deleting the datastore file [datastore1] a16a726b-9673-4de7-9188-19c608a7dc0b {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 858.382784] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-edf95f49-533e-43b6-b519-e85a67f65aa9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.389483] env[61868]: DEBUG oslo_vmware.api [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 858.389483] env[61868]: value = "task-1315465" [ 858.389483] env[61868]: _type = "Task" [ 858.389483] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.398885] env[61868]: DEBUG oslo_vmware.api [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315465, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.422936] env[61868]: DEBUG oslo_vmware.api [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315464, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.437751] env[61868]: DEBUG nova.network.neutron [req-5d944fa5-b95a-4ed5-a353-cec0232f5330 req-3d9f70f1-ca37-46c4-a372-be1c1fde6772 service nova] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Updated VIF entry in instance network info cache for port 548fe1c2-07a5-45a2-b084-cc95216cdcad. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 858.438141] env[61868]: DEBUG nova.network.neutron [req-5d944fa5-b95a-4ed5-a353-cec0232f5330 req-3d9f70f1-ca37-46c4-a372-be1c1fde6772 service nova] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Updating instance_info_cache with network_info: [{"id": "548fe1c2-07a5-45a2-b084-cc95216cdcad", "address": "fa:16:3e:99:ec:d5", "network": {"id": "5b4c5906-b873-4369-855b-e0898ad84812", "bridge": null, "label": "tempest-DeleteServersTestJSON-254512637-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "812e115cc1114d0e8536924c4f5556f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap548fe1c2-07", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.515382] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f6f2c8e8-303e-40f0-b4c0-b789047bfad0 tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Lock "ca85dfd1-e794-41dc-a337-fbd99998e1fa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 104.194s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.658164] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34128e66-53bc-4b20-9a27-53687f64ac5f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.663931] env[61868]: DEBUG nova.network.neutron [req-b6029332-0df1-40d1-be13-2bab39e19071 req-c35b1a84-fbd0-40a6-9986-712ce70a15f0 service nova] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Updated VIF entry in instance network info cache for port 21d6cab8-f691-425c-bfa8-9d2f1c294b64. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 858.664382] env[61868]: DEBUG nova.network.neutron [req-b6029332-0df1-40d1-be13-2bab39e19071 req-c35b1a84-fbd0-40a6-9986-712ce70a15f0 service nova] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Updating instance_info_cache with network_info: [{"id": "21d6cab8-f691-425c-bfa8-9d2f1c294b64", "address": "fa:16:3e:76:a4:70", "network": {"id": "e7457f1a-171a-4c82-a260-1d1a7a95861e", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1032778554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b9b800086764b41a8817d1d8e463a2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21d6cab8-f6", "ovs_interfaceid": "21d6cab8-f691-425c-bfa8-9d2f1c294b64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.668940] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c611633-bbb4-44c6-a991-6ba29827ec87 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.708425] env[61868]: INFO nova.compute.manager [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Rebuilding instance [ 858.711988] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9697e795-9ffd-4385-8123-63bc739938c5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.721081] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f718e9aa-560a-4ed3-be18-4168de6a2010 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.737961] env[61868]: DEBUG nova.compute.provider_tree [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 858.777257] env[61868]: DEBUG nova.compute.manager [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 858.778179] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-496bf371-47a7-4a23-ae6b-33c3d95cac1c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.788184] env[61868]: DEBUG nova.compute.manager [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 858.799906] env[61868]: DEBUG nova.network.neutron [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 858.836831] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315463, 'name': CreateVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.899097] env[61868]: DEBUG oslo_vmware.api [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315465, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.349217} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.899888] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 858.899888] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 858.899888] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 858.923374] env[61868]: DEBUG oslo_vmware.api [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315464, 'name': ReconfigVM_Task, 'duration_secs': 0.919563} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.923893] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Reconfigured VM instance instance-00000042 to attach disk [datastore2] 19c2720c-90bc-47f6-999b-6031f893408d/19c2720c-90bc-47f6-999b-6031f893408d.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 858.924558] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-287fdbba-9dab-49ec-9f3a-1b66cee425b2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.927793] env[61868]: INFO nova.scheduler.client.report [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Deleted allocations for instance a16a726b-9673-4de7-9188-19c608a7dc0b [ 858.931264] env[61868]: DEBUG oslo_vmware.api [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 858.931264] env[61868]: value = "task-1315466" [ 858.931264] env[61868]: _type = "Task" [ 858.931264] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.939587] env[61868]: DEBUG oslo_vmware.api [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315466, 'name': Rename_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.941241] env[61868]: DEBUG oslo_concurrency.lockutils [req-5d944fa5-b95a-4ed5-a353-cec0232f5330 req-3d9f70f1-ca37-46c4-a372-be1c1fde6772 service nova] Releasing lock "refresh_cache-a16a726b-9673-4de7-9188-19c608a7dc0b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.941493] env[61868]: DEBUG nova.compute.manager [req-5d944fa5-b95a-4ed5-a353-cec0232f5330 req-3d9f70f1-ca37-46c4-a372-be1c1fde6772 service nova] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Received event network-changed-29d18263-e917-46dc-8b91-33365c0d3481 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 858.941689] env[61868]: DEBUG nova.compute.manager [req-5d944fa5-b95a-4ed5-a353-cec0232f5330 req-3d9f70f1-ca37-46c4-a372-be1c1fde6772 service nova] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Refreshing instance network info cache due to event network-changed-29d18263-e917-46dc-8b91-33365c0d3481. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 858.941916] env[61868]: DEBUG oslo_concurrency.lockutils [req-5d944fa5-b95a-4ed5-a353-cec0232f5330 req-3d9f70f1-ca37-46c4-a372-be1c1fde6772 service nova] Acquiring lock "refresh_cache-409cad1a-946d-4c58-aa57-1c0bf97fe63f" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.942076] env[61868]: DEBUG oslo_concurrency.lockutils [req-5d944fa5-b95a-4ed5-a353-cec0232f5330 req-3d9f70f1-ca37-46c4-a372-be1c1fde6772 service nova] Acquired lock "refresh_cache-409cad1a-946d-4c58-aa57-1c0bf97fe63f" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.942246] env[61868]: DEBUG nova.network.neutron [req-5d944fa5-b95a-4ed5-a353-cec0232f5330 req-3d9f70f1-ca37-46c4-a372-be1c1fde6772 service nova] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Refreshing network info cache for port 29d18263-e917-46dc-8b91-33365c0d3481 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 859.021908] env[61868]: DEBUG nova.compute.manager [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 859.071554] env[61868]: DEBUG nova.network.neutron [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Updating instance_info_cache with network_info: [{"id": "5159fed6-c58a-4dc0-9b7d-964aa28faaef", "address": "fa:16:3e:9c:1f:04", "network": {"id": "bb1b8c71-26ca-4b8e-8441-de12cef2073f", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-384307023-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "87aa9ce723574a31a0632d3736039ca4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ba27300-88df-4c95-b9e0-a4a8b5039c3c", "external-id": "nsx-vlan-transportzone-681", "segmentation_id": 681, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5159fed6-c5", "ovs_interfaceid": "5159fed6-c58a-4dc0-9b7d-964aa28faaef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.168604] env[61868]: DEBUG oslo_concurrency.lockutils [req-b6029332-0df1-40d1-be13-2bab39e19071 req-c35b1a84-fbd0-40a6-9986-712ce70a15f0 service nova] Releasing lock "refresh_cache-75fdbc11-1ef3-442d-bcf4-55069ba9ab62" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.242535] env[61868]: DEBUG nova.scheduler.client.report [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 859.297088] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 859.297989] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4ec746f5-79c3-4af5-baf6-9d9d7f93a5ee {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.304431] env[61868]: DEBUG oslo_vmware.api [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 859.304431] env[61868]: value = "task-1315467" [ 859.304431] env[61868]: _type = "Task" [ 859.304431] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.312386] env[61868]: DEBUG oslo_vmware.api [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315467, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.337101] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315463, 'name': CreateVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.434086] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.448223] env[61868]: DEBUG oslo_vmware.api [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315466, 'name': Rename_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.546614] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.574248] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Releasing lock "refresh_cache-c6f99cbc-3e81-44b4-b3d1-d8cd821829f2" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.574598] env[61868]: DEBUG nova.compute.manager [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Instance network_info: |[{"id": "5159fed6-c58a-4dc0-9b7d-964aa28faaef", "address": "fa:16:3e:9c:1f:04", "network": {"id": "bb1b8c71-26ca-4b8e-8441-de12cef2073f", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-384307023-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "87aa9ce723574a31a0632d3736039ca4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ba27300-88df-4c95-b9e0-a4a8b5039c3c", "external-id": "nsx-vlan-transportzone-681", "segmentation_id": 681, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5159fed6-c5", "ovs_interfaceid": "5159fed6-c58a-4dc0-9b7d-964aa28faaef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 859.575051] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9c:1f:04', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5ba27300-88df-4c95-b9e0-a4a8b5039c3c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5159fed6-c58a-4dc0-9b7d-964aa28faaef', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 859.582961] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Creating folder: Project (87aa9ce723574a31a0632d3736039ca4). Parent ref: group-v281478. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 859.583257] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-24fbe06d-874f-4ee7-aba7-329b46512687 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.593957] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Created folder: Project (87aa9ce723574a31a0632d3736039ca4) in parent group-v281478. [ 859.594188] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Creating folder: Instances. Parent ref: group-v281535. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 859.594435] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9075dabe-823a-4df4-829a-2efc4d5abe32 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.605985] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Created folder: Instances in parent group-v281535. [ 859.606263] env[61868]: DEBUG oslo.service.loopingcall [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 859.606470] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 859.607139] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fea5b232-1c40-453d-804a-d2bd924c597f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.630513] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0a2a4111-8de4-4541-8ce5-040de2bcae6b tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Acquiring lock "ca85dfd1-e794-41dc-a337-fbd99998e1fa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.630766] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0a2a4111-8de4-4541-8ce5-040de2bcae6b tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Lock "ca85dfd1-e794-41dc-a337-fbd99998e1fa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.630969] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0a2a4111-8de4-4541-8ce5-040de2bcae6b tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Acquiring lock "ca85dfd1-e794-41dc-a337-fbd99998e1fa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.631168] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0a2a4111-8de4-4541-8ce5-040de2bcae6b tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Lock "ca85dfd1-e794-41dc-a337-fbd99998e1fa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.631335] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0a2a4111-8de4-4541-8ce5-040de2bcae6b tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Lock "ca85dfd1-e794-41dc-a337-fbd99998e1fa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.634363] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 859.634363] env[61868]: value = "task-1315470" [ 859.634363] env[61868]: _type = "Task" [ 859.634363] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.634801] env[61868]: INFO nova.compute.manager [None req-0a2a4111-8de4-4541-8ce5-040de2bcae6b tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Terminating instance [ 859.639576] env[61868]: DEBUG nova.compute.manager [None req-0a2a4111-8de4-4541-8ce5-040de2bcae6b tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 859.639576] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0a2a4111-8de4-4541-8ce5-040de2bcae6b tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 859.640553] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5099f86-68a0-43d1-9b3a-1c152e785ee1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.649129] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315470, 'name': CreateVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.651146] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a2a4111-8de4-4541-8ce5-040de2bcae6b tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 859.651376] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b4f4d4be-3554-4007-a988-9b029ab93132 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.657095] env[61868]: DEBUG oslo_vmware.api [None req-0a2a4111-8de4-4541-8ce5-040de2bcae6b tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Waiting for the task: (returnval){ [ 859.657095] env[61868]: value = "task-1315471" [ 859.657095] env[61868]: _type = "Task" [ 859.657095] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.665066] env[61868]: DEBUG oslo_vmware.api [None req-0a2a4111-8de4-4541-8ce5-040de2bcae6b tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Task: {'id': task-1315471, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.749364] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.971s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.754679] env[61868]: DEBUG oslo_concurrency.lockutils [None req-fd684283-1baf-4fb4-90ff-bd8128b2a028 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.972s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.755152] env[61868]: DEBUG nova.objects.instance [None req-fd684283-1baf-4fb4-90ff-bd8128b2a028 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Lazy-loading 'resources' on Instance uuid 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 859.766195] env[61868]: DEBUG nova.network.neutron [req-5d944fa5-b95a-4ed5-a353-cec0232f5330 req-3d9f70f1-ca37-46c4-a372-be1c1fde6772 service nova] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Updated VIF entry in instance network info cache for port 29d18263-e917-46dc-8b91-33365c0d3481. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 859.766620] env[61868]: DEBUG nova.network.neutron [req-5d944fa5-b95a-4ed5-a353-cec0232f5330 req-3d9f70f1-ca37-46c4-a372-be1c1fde6772 service nova] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Updating instance_info_cache with network_info: [{"id": "29d18263-e917-46dc-8b91-33365c0d3481", "address": "fa:16:3e:c6:07:cd", "network": {"id": "0fa2f8e1-ee57-450f-9d06-9a056566585b", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-729551994-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7471600a510e4feeb0c819d0cca5eb48", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap29d18263-e9", "ovs_interfaceid": "29d18263-e917-46dc-8b91-33365c0d3481", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.782626] env[61868]: INFO nova.scheduler.client.report [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Deleted allocations for instance 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02 [ 859.799593] env[61868]: DEBUG nova.compute.manager [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 859.814530] env[61868]: DEBUG oslo_vmware.api [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315467, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.825015] env[61868]: DEBUG nova.virt.hardware [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 859.825343] env[61868]: DEBUG nova.virt.hardware [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 859.825535] env[61868]: DEBUG nova.virt.hardware [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 859.825788] env[61868]: DEBUG nova.virt.hardware [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 859.826028] env[61868]: DEBUG nova.virt.hardware [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 859.826219] env[61868]: DEBUG nova.virt.hardware [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 859.826452] env[61868]: DEBUG nova.virt.hardware [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 859.826626] env[61868]: DEBUG nova.virt.hardware [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 859.826820] env[61868]: DEBUG nova.virt.hardware [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 859.826999] env[61868]: DEBUG nova.virt.hardware [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 859.827204] env[61868]: DEBUG nova.virt.hardware [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 859.828084] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33dd2bde-973d-4c13-a91e-21c0ae6363f0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.833293] env[61868]: DEBUG nova.compute.manager [req-a064489f-6dc0-4386-8278-c0413556b52b req-b1a490c2-fadb-41af-a86c-de5d071529d2 service nova] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Received event network-vif-plugged-5159fed6-c58a-4dc0-9b7d-964aa28faaef {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 859.833470] env[61868]: DEBUG oslo_concurrency.lockutils [req-a064489f-6dc0-4386-8278-c0413556b52b req-b1a490c2-fadb-41af-a86c-de5d071529d2 service nova] Acquiring lock "c6f99cbc-3e81-44b4-b3d1-d8cd821829f2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.833693] env[61868]: DEBUG oslo_concurrency.lockutils [req-a064489f-6dc0-4386-8278-c0413556b52b req-b1a490c2-fadb-41af-a86c-de5d071529d2 service nova] Lock "c6f99cbc-3e81-44b4-b3d1-d8cd821829f2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.833894] env[61868]: DEBUG oslo_concurrency.lockutils [req-a064489f-6dc0-4386-8278-c0413556b52b req-b1a490c2-fadb-41af-a86c-de5d071529d2 service nova] Lock "c6f99cbc-3e81-44b4-b3d1-d8cd821829f2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.834103] env[61868]: DEBUG nova.compute.manager [req-a064489f-6dc0-4386-8278-c0413556b52b req-b1a490c2-fadb-41af-a86c-de5d071529d2 service nova] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] No waiting events found dispatching network-vif-plugged-5159fed6-c58a-4dc0-9b7d-964aa28faaef {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 859.834381] env[61868]: WARNING nova.compute.manager [req-a064489f-6dc0-4386-8278-c0413556b52b req-b1a490c2-fadb-41af-a86c-de5d071529d2 service nova] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Received unexpected event network-vif-plugged-5159fed6-c58a-4dc0-9b7d-964aa28faaef for instance with vm_state building and task_state spawning. [ 859.834572] env[61868]: DEBUG nova.compute.manager [req-a064489f-6dc0-4386-8278-c0413556b52b req-b1a490c2-fadb-41af-a86c-de5d071529d2 service nova] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Received event network-changed-5159fed6-c58a-4dc0-9b7d-964aa28faaef {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 859.834770] env[61868]: DEBUG nova.compute.manager [req-a064489f-6dc0-4386-8278-c0413556b52b req-b1a490c2-fadb-41af-a86c-de5d071529d2 service nova] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Refreshing instance network info cache due to event network-changed-5159fed6-c58a-4dc0-9b7d-964aa28faaef. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 859.835039] env[61868]: DEBUG oslo_concurrency.lockutils [req-a064489f-6dc0-4386-8278-c0413556b52b req-b1a490c2-fadb-41af-a86c-de5d071529d2 service nova] Acquiring lock "refresh_cache-c6f99cbc-3e81-44b4-b3d1-d8cd821829f2" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.835199] env[61868]: DEBUG oslo_concurrency.lockutils [req-a064489f-6dc0-4386-8278-c0413556b52b req-b1a490c2-fadb-41af-a86c-de5d071529d2 service nova] Acquired lock "refresh_cache-c6f99cbc-3e81-44b4-b3d1-d8cd821829f2" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.835412] env[61868]: DEBUG nova.network.neutron [req-a064489f-6dc0-4386-8278-c0413556b52b req-b1a490c2-fadb-41af-a86c-de5d071529d2 service nova] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Refreshing network info cache for port 5159fed6-c58a-4dc0-9b7d-964aa28faaef {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 859.851451] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c8e7c39-0345-46cd-b9f1-10750535090d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.855560] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315463, 'name': CreateVM_Task, 'duration_secs': 1.576869} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.856272] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 859.856773] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.856987] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.857518] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 859.857816] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-68c134d5-4395-4778-aebb-46887404b5a2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.867989] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Instance VIF info [] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 859.876926] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Creating folder: Project (b343a79b52a04458b7fdc1ad9cde9508). Parent ref: group-v281478. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 859.877844] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e058809d-137c-4403-a7c7-e6aeef85f86f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.880822] env[61868]: DEBUG oslo_vmware.api [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the task: (returnval){ [ 859.880822] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]525430ff-c8c6-61ac-5b68-fd5604647c0e" [ 859.880822] env[61868]: _type = "Task" [ 859.880822] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.889411] env[61868]: DEBUG oslo_vmware.api [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]525430ff-c8c6-61ac-5b68-fd5604647c0e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.890605] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Created folder: Project (b343a79b52a04458b7fdc1ad9cde9508) in parent group-v281478. [ 859.890836] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Creating folder: Instances. Parent ref: group-v281538. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 859.891095] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-31039f55-536e-4d24-8bfa-2a38b90e0600 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.899727] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Created folder: Instances in parent group-v281538. [ 859.899983] env[61868]: DEBUG oslo.service.loopingcall [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 859.900216] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 859.900425] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8fc42b13-f669-49a4-8714-8bf68f2fc8a4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.916851] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 859.916851] env[61868]: value = "task-1315474" [ 859.916851] env[61868]: _type = "Task" [ 859.916851] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.924894] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315474, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.944765] env[61868]: DEBUG oslo_vmware.api [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315466, 'name': Rename_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.128406] env[61868]: DEBUG oslo_concurrency.lockutils [None req-75a27b94-21b6-4932-bd97-f47ebe7e57ec tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "a16a726b-9673-4de7-9188-19c608a7dc0b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.145713] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315470, 'name': CreateVM_Task, 'duration_secs': 0.464109} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.145920] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 860.146647] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.146857] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.147392] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 860.147492] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aea3338a-a037-46de-89a5-6af133b72f29 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.151990] env[61868]: DEBUG oslo_vmware.api [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Waiting for the task: (returnval){ [ 860.151990] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]529bf196-3085-4565-efd8-f269b91918e3" [ 860.151990] env[61868]: _type = "Task" [ 860.151990] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.160178] env[61868]: DEBUG oslo_vmware.api [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]529bf196-3085-4565-efd8-f269b91918e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.167224] env[61868]: DEBUG oslo_vmware.api [None req-0a2a4111-8de4-4541-8ce5-040de2bcae6b tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Task: {'id': task-1315471, 'name': PowerOffVM_Task, 'duration_secs': 0.202151} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.167468] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a2a4111-8de4-4541-8ce5-040de2bcae6b tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 860.167640] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0a2a4111-8de4-4541-8ce5-040de2bcae6b tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 860.167863] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ec3bc03c-0da7-444f-8a8c-301958630bcd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.226439] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0a2a4111-8de4-4541-8ce5-040de2bcae6b tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 860.226697] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0a2a4111-8de4-4541-8ce5-040de2bcae6b tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 860.226886] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a2a4111-8de4-4541-8ce5-040de2bcae6b tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Deleting the datastore file [datastore1] ca85dfd1-e794-41dc-a337-fbd99998e1fa {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 860.227169] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-249ca780-7544-4f3b-893b-1923a47747ac {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.233990] env[61868]: DEBUG oslo_vmware.api [None req-0a2a4111-8de4-4541-8ce5-040de2bcae6b tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Waiting for the task: (returnval){ [ 860.233990] env[61868]: value = "task-1315476" [ 860.233990] env[61868]: _type = "Task" [ 860.233990] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.241857] env[61868]: DEBUG oslo_vmware.api [None req-0a2a4111-8de4-4541-8ce5-040de2bcae6b tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Task: {'id': task-1315476, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.270726] env[61868]: DEBUG oslo_concurrency.lockutils [req-5d944fa5-b95a-4ed5-a353-cec0232f5330 req-3d9f70f1-ca37-46c4-a372-be1c1fde6772 service nova] Releasing lock "refresh_cache-409cad1a-946d-4c58-aa57-1c0bf97fe63f" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.294924] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e3ce8e10-2ffd-44c4-950e-906da31d184b tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Lock "02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.056s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.318693] env[61868]: DEBUG oslo_vmware.api [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315467, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.392762] env[61868]: DEBUG oslo_vmware.api [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]525430ff-c8c6-61ac-5b68-fd5604647c0e, 'name': SearchDatastore_Task, 'duration_secs': 0.016837} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.393074] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.393315] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 860.393551] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.393701] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.393908] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 860.394191] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f166be4a-1c55-4642-b920-86dd02e61a8b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.405064] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 860.405298] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 860.408916] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-62cfe6c5-f1c8-4738-b99a-8b38b99f5096 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.416263] env[61868]: DEBUG oslo_vmware.api [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the task: (returnval){ [ 860.416263] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52d42856-c3b3-d255-7692-d1bdead73231" [ 860.416263] env[61868]: _type = "Task" [ 860.416263] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.428303] env[61868]: DEBUG oslo_vmware.api [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52d42856-c3b3-d255-7692-d1bdead73231, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.433635] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315474, 'name': CreateVM_Task, 'duration_secs': 0.452918} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.433841] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 860.434280] env[61868]: DEBUG oslo_concurrency.lockutils [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.434449] env[61868]: DEBUG oslo_concurrency.lockutils [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.434760] env[61868]: DEBUG oslo_concurrency.lockutils [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 860.435073] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2e29d40b-e27c-4170-8145-ffee951b0fd1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.443844] env[61868]: DEBUG oslo_vmware.api [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Waiting for the task: (returnval){ [ 860.443844] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52967ead-59f7-7a94-3d5a-31628314f82b" [ 860.443844] env[61868]: _type = "Task" [ 860.443844] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.447151] env[61868]: DEBUG oslo_vmware.api [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315466, 'name': Rename_Task, 'duration_secs': 1.401997} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.450068] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 860.452622] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-770f0622-cd59-41e4-8209-ac84545f8e8f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.461376] env[61868]: DEBUG oslo_vmware.api [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52967ead-59f7-7a94-3d5a-31628314f82b, 'name': SearchDatastore_Task, 'duration_secs': 0.011548} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.462368] env[61868]: DEBUG oslo_concurrency.lockutils [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.462527] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 860.462986] env[61868]: DEBUG oslo_concurrency.lockutils [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.463069] env[61868]: DEBUG oslo_vmware.api [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 860.463069] env[61868]: value = "task-1315477" [ 860.463069] env[61868]: _type = "Task" [ 860.463069] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.471901] env[61868]: DEBUG oslo_vmware.api [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315477, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.495443] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Acquiring lock "d35953d8-692d-498e-baf5-96ef381ce12b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.496548] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Lock "d35953d8-692d-498e-baf5-96ef381ce12b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.496548] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Acquiring lock "d35953d8-692d-498e-baf5-96ef381ce12b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.496548] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Lock "d35953d8-692d-498e-baf5-96ef381ce12b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.496548] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Lock "d35953d8-692d-498e-baf5-96ef381ce12b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.499869] env[61868]: INFO nova.compute.manager [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Terminating instance [ 860.503550] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Acquiring lock "refresh_cache-d35953d8-692d-498e-baf5-96ef381ce12b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.503759] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Acquired lock "refresh_cache-d35953d8-692d-498e-baf5-96ef381ce12b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.503961] env[61868]: DEBUG nova.network.neutron [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 860.581204] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-312d249e-85a8-4aa2-b04d-47ee5e226eb4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.589274] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e4f8bfa-76d5-4fb0-9b7b-5d949face74e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.623675] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee1d75e0-eb87-444a-96e1-af3a01657add {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.631336] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9a1fefb-9c20-42e2-a9ea-18dd44a9c5f3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.635701] env[61868]: DEBUG nova.network.neutron [req-a064489f-6dc0-4386-8278-c0413556b52b req-b1a490c2-fadb-41af-a86c-de5d071529d2 service nova] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Updated VIF entry in instance network info cache for port 5159fed6-c58a-4dc0-9b7d-964aa28faaef. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 860.636044] env[61868]: DEBUG nova.network.neutron [req-a064489f-6dc0-4386-8278-c0413556b52b req-b1a490c2-fadb-41af-a86c-de5d071529d2 service nova] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Updating instance_info_cache with network_info: [{"id": "5159fed6-c58a-4dc0-9b7d-964aa28faaef", "address": "fa:16:3e:9c:1f:04", "network": {"id": "bb1b8c71-26ca-4b8e-8441-de12cef2073f", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-384307023-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "87aa9ce723574a31a0632d3736039ca4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ba27300-88df-4c95-b9e0-a4a8b5039c3c", "external-id": "nsx-vlan-transportzone-681", "segmentation_id": 681, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5159fed6-c5", "ovs_interfaceid": "5159fed6-c58a-4dc0-9b7d-964aa28faaef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.646860] env[61868]: DEBUG nova.compute.provider_tree [None req-fd684283-1baf-4fb4-90ff-bd8128b2a028 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 860.660953] env[61868]: DEBUG oslo_vmware.api [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]529bf196-3085-4565-efd8-f269b91918e3, 'name': SearchDatastore_Task, 'duration_secs': 0.010034} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.661277] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.661509] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 860.661772] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.661930] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.662127] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 860.662376] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cebe8f26-5d25-4582-b647-c27bd4d9edd8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.670250] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 860.670428] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 860.671106] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f072e3ab-91f8-4284-a700-d858038568f0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.675761] env[61868]: DEBUG oslo_vmware.api [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Waiting for the task: (returnval){ [ 860.675761] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]528ad2e9-ec35-c426-1333-2514d9a8e881" [ 860.675761] env[61868]: _type = "Task" [ 860.675761] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.682852] env[61868]: DEBUG oslo_vmware.api [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]528ad2e9-ec35-c426-1333-2514d9a8e881, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.744615] env[61868]: DEBUG oslo_vmware.api [None req-0a2a4111-8de4-4541-8ce5-040de2bcae6b tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Task: {'id': task-1315476, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.177082} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.744835] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a2a4111-8de4-4541-8ce5-040de2bcae6b tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 860.745285] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0a2a4111-8de4-4541-8ce5-040de2bcae6b tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 860.745495] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0a2a4111-8de4-4541-8ce5-040de2bcae6b tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 860.745681] env[61868]: INFO nova.compute.manager [None req-0a2a4111-8de4-4541-8ce5-040de2bcae6b tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Took 1.11 seconds to destroy the instance on the hypervisor. [ 860.745926] env[61868]: DEBUG oslo.service.loopingcall [None req-0a2a4111-8de4-4541-8ce5-040de2bcae6b tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 860.746141] env[61868]: DEBUG nova.compute.manager [-] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 860.746236] env[61868]: DEBUG nova.network.neutron [-] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 860.815557] env[61868]: DEBUG oslo_vmware.api [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315467, 'name': PowerOffVM_Task, 'duration_secs': 1.161678} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.815827] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 860.816062] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 860.816812] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-371c92a0-6a93-4a23-8cdc-5b51783ffcfe {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.823389] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 860.823622] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-84de3ed6-f32d-4e4d-8aa6-55852f9c4baf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.893024] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 860.893143] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Deleting contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 860.893330] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Deleting the datastore file [datastore2] c7b10943-5136-44c8-b4e7-59651220a333 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 860.893602] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-03639a95-eb62-405b-bb63-a187209853d6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.899818] env[61868]: DEBUG oslo_vmware.api [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 860.899818] env[61868]: value = "task-1315479" [ 860.899818] env[61868]: _type = "Task" [ 860.899818] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.908574] env[61868]: DEBUG oslo_vmware.api [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315479, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.925750] env[61868]: DEBUG oslo_vmware.api [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52d42856-c3b3-d255-7692-d1bdead73231, 'name': SearchDatastore_Task, 'duration_secs': 0.011862} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.927139] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-19e12533-d82c-463a-a499-0389b79d7e30 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.931177] env[61868]: DEBUG oslo_vmware.api [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the task: (returnval){ [ 860.931177] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5226bb78-2b4c-3b05-cd15-ee8f5ac51ebf" [ 860.931177] env[61868]: _type = "Task" [ 860.931177] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.939450] env[61868]: DEBUG oslo_vmware.api [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5226bb78-2b4c-3b05-cd15-ee8f5ac51ebf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.972286] env[61868]: DEBUG oslo_vmware.api [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315477, 'name': PowerOnVM_Task, 'duration_secs': 0.430874} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.972555] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 860.972819] env[61868]: INFO nova.compute.manager [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Took 9.59 seconds to spawn the instance on the hypervisor. [ 860.973042] env[61868]: DEBUG nova.compute.manager [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 860.973801] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d766da4-c321-45cf-9076-1275cf094f35 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.028993] env[61868]: DEBUG nova.network.neutron [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 861.084816] env[61868]: DEBUG nova.network.neutron [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.138771] env[61868]: DEBUG oslo_concurrency.lockutils [req-a064489f-6dc0-4386-8278-c0413556b52b req-b1a490c2-fadb-41af-a86c-de5d071529d2 service nova] Releasing lock "refresh_cache-c6f99cbc-3e81-44b4-b3d1-d8cd821829f2" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.150193] env[61868]: DEBUG nova.scheduler.client.report [None req-fd684283-1baf-4fb4-90ff-bd8128b2a028 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 861.184576] env[61868]: DEBUG nova.compute.manager [req-a5725026-ddb6-42fe-bc61-c97f631b2b2a req-9b423491-7b22-4c39-92af-535a07be397a service nova] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Received event network-vif-deleted-9196ece9-3dc0-411e-b759-7e201fa02c56 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 861.184718] env[61868]: INFO nova.compute.manager [req-a5725026-ddb6-42fe-bc61-c97f631b2b2a req-9b423491-7b22-4c39-92af-535a07be397a service nova] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Neutron deleted interface 9196ece9-3dc0-411e-b759-7e201fa02c56; detaching it from the instance and deleting it from the info cache [ 861.184897] env[61868]: DEBUG nova.network.neutron [req-a5725026-ddb6-42fe-bc61-c97f631b2b2a req-9b423491-7b22-4c39-92af-535a07be397a service nova] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.190274] env[61868]: DEBUG oslo_vmware.api [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]528ad2e9-ec35-c426-1333-2514d9a8e881, 'name': SearchDatastore_Task, 'duration_secs': 0.022436} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.191035] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2997b5c-b10f-4b4f-8c52-e3ce37aed577 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.196788] env[61868]: DEBUG oslo_vmware.api [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Waiting for the task: (returnval){ [ 861.196788] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]529439c4-38bd-4657-4db1-11795847fb7c" [ 861.196788] env[61868]: _type = "Task" [ 861.196788] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.204824] env[61868]: DEBUG oslo_vmware.api [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]529439c4-38bd-4657-4db1-11795847fb7c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.410034] env[61868]: DEBUG oslo_vmware.api [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315479, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.405587} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.410157] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 861.410322] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Deleted contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 861.410496] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 861.441508] env[61868]: DEBUG oslo_vmware.api [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5226bb78-2b4c-3b05-cd15-ee8f5ac51ebf, 'name': SearchDatastore_Task, 'duration_secs': 0.010281} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.442462] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.442699] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 75fdbc11-1ef3-442d-bcf4-55069ba9ab62/75fdbc11-1ef3-442d-bcf4-55069ba9ab62.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 861.443015] env[61868]: DEBUG oslo_concurrency.lockutils [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.443229] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 861.443445] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cd39b1a3-08c6-43e6-8ea8-ed89f961a2bb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.445547] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2dae8e1f-d11d-4586-8ea6-bd9a6235ebc6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.452057] env[61868]: DEBUG oslo_vmware.api [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the task: (returnval){ [ 861.452057] env[61868]: value = "task-1315480" [ 861.452057] env[61868]: _type = "Task" [ 861.452057] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.456157] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 861.456364] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 861.457330] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-50c6928c-a0c1-42d6-92c5-12f016080601 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.462439] env[61868]: DEBUG oslo_vmware.api [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315480, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.465185] env[61868]: DEBUG oslo_vmware.api [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Waiting for the task: (returnval){ [ 861.465185] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]524217fe-7edd-6ca2-917c-cd0c98bbf299" [ 861.465185] env[61868]: _type = "Task" [ 861.465185] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.472577] env[61868]: DEBUG oslo_vmware.api [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]524217fe-7edd-6ca2-917c-cd0c98bbf299, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.491817] env[61868]: INFO nova.compute.manager [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Took 22.42 seconds to build instance. [ 861.587124] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Releasing lock "refresh_cache-d35953d8-692d-498e-baf5-96ef381ce12b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.587619] env[61868]: DEBUG nova.compute.manager [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 861.587821] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 861.588717] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8db173ed-41fb-449a-9fbd-826826f30746 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.596301] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 861.596523] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9412b98c-e0ca-4d9b-b115-9a0fc206bbdd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.602267] env[61868]: DEBUG oslo_vmware.api [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Waiting for the task: (returnval){ [ 861.602267] env[61868]: value = "task-1315481" [ 861.602267] env[61868]: _type = "Task" [ 861.602267] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.609767] env[61868]: DEBUG oslo_vmware.api [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315481, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.627631] env[61868]: DEBUG nova.network.neutron [-] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.659497] env[61868]: DEBUG oslo_concurrency.lockutils [None req-fd684283-1baf-4fb4-90ff-bd8128b2a028 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.905s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.661722] env[61868]: DEBUG oslo_concurrency.lockutils [None req-05a0085d-ae8b-4e73-9106-330b108967d5 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.838s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.662507] env[61868]: DEBUG nova.objects.instance [None req-05a0085d-ae8b-4e73-9106-330b108967d5 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Lazy-loading 'resources' on Instance uuid 7788f9f8-c08e-46f8-b204-070215223436 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 861.684462] env[61868]: INFO nova.scheduler.client.report [None req-fd684283-1baf-4fb4-90ff-bd8128b2a028 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Deleted allocations for instance 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc [ 861.689609] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b3461258-3685-4349-99ce-898278ba0836 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.699474] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c0741f5-b9fd-49c9-9b22-7a599dd1ee01 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.725066] env[61868]: DEBUG oslo_vmware.api [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]529439c4-38bd-4657-4db1-11795847fb7c, 'name': SearchDatastore_Task, 'duration_secs': 0.009615} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.733897] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.734238] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] c6f99cbc-3e81-44b4-b3d1-d8cd821829f2/c6f99cbc-3e81-44b4-b3d1-d8cd821829f2.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 861.734609] env[61868]: DEBUG nova.compute.manager [req-a5725026-ddb6-42fe-bc61-c97f631b2b2a req-9b423491-7b22-4c39-92af-535a07be397a service nova] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Detach interface failed, port_id=9196ece9-3dc0-411e-b759-7e201fa02c56, reason: Instance ca85dfd1-e794-41dc-a337-fbd99998e1fa could not be found. {{(pid=61868) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 861.735623] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7df6944e-2fd8-4ae0-989d-393bede33011 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.742108] env[61868]: DEBUG oslo_vmware.api [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Waiting for the task: (returnval){ [ 861.742108] env[61868]: value = "task-1315482" [ 861.742108] env[61868]: _type = "Task" [ 861.742108] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.749892] env[61868]: DEBUG oslo_vmware.api [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Task: {'id': task-1315482, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.868369] env[61868]: INFO nova.compute.manager [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Rescuing [ 861.868689] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Acquiring lock "refresh_cache-19c2720c-90bc-47f6-999b-6031f893408d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.868893] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Acquired lock "refresh_cache-19c2720c-90bc-47f6-999b-6031f893408d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.869169] env[61868]: DEBUG nova.network.neutron [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 861.967150] env[61868]: DEBUG oslo_vmware.api [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315480, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.981223] env[61868]: DEBUG oslo_vmware.api [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]524217fe-7edd-6ca2-917c-cd0c98bbf299, 'name': SearchDatastore_Task, 'duration_secs': 0.011213} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.982871] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60d0aa80-7121-49a7-a7c6-e7996e1ff53c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.988861] env[61868]: DEBUG oslo_vmware.api [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Waiting for the task: (returnval){ [ 861.988861] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52480366-7f35-fbad-44b2-dc1b5a8dd44a" [ 861.988861] env[61868]: _type = "Task" [ 861.988861] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.994839] env[61868]: DEBUG oslo_concurrency.lockutils [None req-44067d7f-c4d9-4615-8225-a02bcb790fa8 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Lock "19c2720c-90bc-47f6-999b-6031f893408d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 106.025s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.000582] env[61868]: DEBUG oslo_vmware.api [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52480366-7f35-fbad-44b2-dc1b5a8dd44a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.112244] env[61868]: DEBUG oslo_vmware.api [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315481, 'name': PowerOffVM_Task, 'duration_secs': 0.123958} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.112510] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 862.112688] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 862.112978] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-722c97c5-4c58-4cdb-968c-ba80137cb0bf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.131061] env[61868]: INFO nova.compute.manager [-] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Took 1.38 seconds to deallocate network for instance. [ 862.158274] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 862.158274] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 862.158274] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Deleting the datastore file [datastore1] d35953d8-692d-498e-baf5-96ef381ce12b {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 862.158500] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a895da8d-d478-46f1-bf89-031839347182 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.170069] env[61868]: DEBUG oslo_vmware.api [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Waiting for the task: (returnval){ [ 862.170069] env[61868]: value = "task-1315484" [ 862.170069] env[61868]: _type = "Task" [ 862.170069] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.184841] env[61868]: DEBUG oslo_vmware.api [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315484, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.194029] env[61868]: DEBUG oslo_concurrency.lockutils [None req-fd684283-1baf-4fb4-90ff-bd8128b2a028 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Lock "0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.546s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.252017] env[61868]: DEBUG oslo_vmware.api [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Task: {'id': task-1315482, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.453663] env[61868]: DEBUG nova.virt.hardware [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 862.453935] env[61868]: DEBUG nova.virt.hardware [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 862.454119] env[61868]: DEBUG nova.virt.hardware [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 862.454318] env[61868]: DEBUG nova.virt.hardware [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 862.454474] env[61868]: DEBUG nova.virt.hardware [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 862.454627] env[61868]: DEBUG nova.virt.hardware [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 862.454840] env[61868]: DEBUG nova.virt.hardware [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 862.455014] env[61868]: DEBUG nova.virt.hardware [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 862.455291] env[61868]: DEBUG nova.virt.hardware [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 862.455470] env[61868]: DEBUG nova.virt.hardware [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 862.455653] env[61868]: DEBUG nova.virt.hardware [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 862.457124] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bebefe7-bae3-4889-977a-4560a3f9b631 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.472452] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0bd3e64-0251-423a-9602-c98756f17cc5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.476366] env[61868]: DEBUG oslo_vmware.api [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315480, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.704506} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.478795] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 75fdbc11-1ef3-442d-bcf4-55069ba9ab62/75fdbc11-1ef3-442d-bcf4-55069ba9ab62.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 862.479037] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 862.479756] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-176be56c-826b-4c8c-b050-21d4c4db454d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.490136] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:63:08:8e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24b91e3b-50f1-4a16-b929-942e6b31b2d8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7b94f787-9d7e-4edf-92b0-44d37a7e93ec', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 862.497664] env[61868]: DEBUG oslo.service.loopingcall [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 862.501841] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 862.502582] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3925f33b-426a-4af7-85a7-fed05964dc5f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.506462] env[61868]: DEBUG nova.compute.manager [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 862.508892] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5f5b1910-415b-47e1-bd85-8cd2a27f8716 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.523269] env[61868]: DEBUG oslo_vmware.api [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the task: (returnval){ [ 862.523269] env[61868]: value = "task-1315485" [ 862.523269] env[61868]: _type = "Task" [ 862.523269] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.536887] env[61868]: DEBUG oslo_vmware.api [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52480366-7f35-fbad-44b2-dc1b5a8dd44a, 'name': SearchDatastore_Task, 'duration_secs': 0.067418} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.537465] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 862.537465] env[61868]: value = "task-1315486" [ 862.537465] env[61868]: _type = "Task" [ 862.537465] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.538071] env[61868]: DEBUG oslo_concurrency.lockutils [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.538365] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] cf063dbc-7b5a-4836-91bf-a0aa33cca6bc/cf063dbc-7b5a-4836-91bf-a0aa33cca6bc.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 862.539380] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-973b41e4-57ce-4128-8b34-f7ff3efbfcef {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.545324] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-30b6bd47-1e94-4165-bab5-f2c43c07cd13 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.549673] env[61868]: DEBUG oslo_vmware.api [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315485, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.559016] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315486, 'name': CreateVM_Task} progress is 15%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.586428] env[61868]: DEBUG oslo_vmware.api [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Waiting for the task: (returnval){ [ 862.586428] env[61868]: value = "task-1315487" [ 862.586428] env[61868]: _type = "Task" [ 862.586428] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.587146] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-883dad9b-a765-4063-b5a6-37c8c8d37c4a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.599517] env[61868]: DEBUG oslo_vmware.api [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Task: {'id': task-1315487, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.600819] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e2f126a-74ab-40f3-9d1c-12ead94561cc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.615068] env[61868]: DEBUG nova.compute.provider_tree [None req-05a0085d-ae8b-4e73-9106-330b108967d5 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 862.639054] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0a2a4111-8de4-4541-8ce5-040de2bcae6b tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.680221] env[61868]: DEBUG oslo_vmware.api [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Task: {'id': task-1315484, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.312127} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.682664] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 862.682889] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 862.683131] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 862.683351] env[61868]: INFO nova.compute.manager [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Took 1.10 seconds to destroy the instance on the hypervisor. [ 862.683640] env[61868]: DEBUG oslo.service.loopingcall [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 862.683882] env[61868]: DEBUG nova.compute.manager [-] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 862.684014] env[61868]: DEBUG nova.network.neutron [-] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 862.698787] env[61868]: DEBUG nova.network.neutron [-] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 862.754848] env[61868]: DEBUG oslo_vmware.api [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Task: {'id': task-1315482, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.736473} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.755336] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] c6f99cbc-3e81-44b4-b3d1-d8cd821829f2/c6f99cbc-3e81-44b4-b3d1-d8cd821829f2.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 862.755750] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 862.756149] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c6aaa4f7-f02a-492c-ba4b-9bdbd4bda93d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.762331] env[61868]: DEBUG oslo_vmware.api [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Waiting for the task: (returnval){ [ 862.762331] env[61868]: value = "task-1315488" [ 862.762331] env[61868]: _type = "Task" [ 862.762331] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.765935] env[61868]: INFO nova.compute.manager [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Rebuilding instance [ 862.772649] env[61868]: DEBUG oslo_vmware.api [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Task: {'id': task-1315488, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.813963] env[61868]: DEBUG nova.compute.manager [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 862.817028] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aa4bfde-b316-49f7-aef3-656af2e0a842 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.828490] env[61868]: DEBUG nova.network.neutron [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Updating instance_info_cache with network_info: [{"id": "77081d78-be71-47a0-a7be-8511ae0243d4", "address": "fa:16:3e:79:ff:17", "network": {"id": "094eaf15-c932-4d4f-8736-fe5239160507", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1905020504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "7541757d43c74b93acf728aa2fb0f425", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap77081d78-be", "ovs_interfaceid": "77081d78-be71-47a0-a7be-8511ae0243d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.058736] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315486, 'name': CreateVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.060539] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.064224] env[61868]: DEBUG oslo_vmware.api [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315485, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072738} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.064545] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 863.065405] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f604cc53-c95c-4453-958f-73371666aefc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.094104] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 75fdbc11-1ef3-442d-bcf4-55069ba9ab62/75fdbc11-1ef3-442d-bcf4-55069ba9ab62.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 863.094488] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8db3e7d1-a635-4bb1-b961-03a116da6ce0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.117859] env[61868]: DEBUG nova.scheduler.client.report [None req-05a0085d-ae8b-4e73-9106-330b108967d5 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 863.130891] env[61868]: DEBUG oslo_vmware.api [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the task: (returnval){ [ 863.130891] env[61868]: value = "task-1315489" [ 863.130891] env[61868]: _type = "Task" [ 863.130891] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.131071] env[61868]: DEBUG oslo_vmware.api [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Task: {'id': task-1315487, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.140610] env[61868]: DEBUG oslo_vmware.api [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315489, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.201161] env[61868]: DEBUG nova.network.neutron [-] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.272763] env[61868]: DEBUG oslo_vmware.api [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Task: {'id': task-1315488, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065376} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.273046] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 863.273934] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b51c1d0-7985-492b-a4e3-b529850f4469 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.298787] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] c6f99cbc-3e81-44b4-b3d1-d8cd821829f2/c6f99cbc-3e81-44b4-b3d1-d8cd821829f2.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 863.299550] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-06281cbd-3e60-43db-92d9-221506499258 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.319012] env[61868]: DEBUG oslo_vmware.api [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Waiting for the task: (returnval){ [ 863.319012] env[61868]: value = "task-1315490" [ 863.319012] env[61868]: _type = "Task" [ 863.319012] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.328703] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 863.329022] env[61868]: DEBUG oslo_vmware.api [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Task: {'id': task-1315490, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.329251] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5c872506-a5c8-4300-b4cb-f246b59f77c5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.331041] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Releasing lock "refresh_cache-19c2720c-90bc-47f6-999b-6031f893408d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.338083] env[61868]: DEBUG oslo_vmware.api [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Waiting for the task: (returnval){ [ 863.338083] env[61868]: value = "task-1315491" [ 863.338083] env[61868]: _type = "Task" [ 863.338083] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.346807] env[61868]: DEBUG oslo_vmware.api [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Task: {'id': task-1315491, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.553337] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315486, 'name': CreateVM_Task, 'duration_secs': 0.663668} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.553528] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 863.554278] env[61868]: DEBUG oslo_concurrency.lockutils [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.554449] env[61868]: DEBUG oslo_concurrency.lockutils [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.554792] env[61868]: DEBUG oslo_concurrency.lockutils [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 863.555086] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a43d2ec-da59-4366-8a5d-dc93129c6552 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.559966] env[61868]: DEBUG oslo_vmware.api [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 863.559966] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e00042-f598-98db-1bcb-7fc0826019c2" [ 863.559966] env[61868]: _type = "Task" [ 863.559966] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.568272] env[61868]: DEBUG oslo_vmware.api [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e00042-f598-98db-1bcb-7fc0826019c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.604793] env[61868]: DEBUG oslo_vmware.api [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Task: {'id': task-1315487, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.751958} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.605160] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] cf063dbc-7b5a-4836-91bf-a0aa33cca6bc/cf063dbc-7b5a-4836-91bf-a0aa33cca6bc.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 863.605301] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 863.605580] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3efd04eb-a4df-4a6b-8ea7-7d6f82d073d3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.614314] env[61868]: DEBUG oslo_vmware.api [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Waiting for the task: (returnval){ [ 863.614314] env[61868]: value = "task-1315492" [ 863.614314] env[61868]: _type = "Task" [ 863.614314] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.624011] env[61868]: DEBUG oslo_vmware.api [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Task: {'id': task-1315492, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.632534] env[61868]: DEBUG oslo_concurrency.lockutils [None req-05a0085d-ae8b-4e73-9106-330b108967d5 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.971s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.635424] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.706s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.636695] env[61868]: INFO nova.compute.claims [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 863.648812] env[61868]: DEBUG oslo_vmware.api [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315489, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.656745] env[61868]: INFO nova.scheduler.client.report [None req-05a0085d-ae8b-4e73-9106-330b108967d5 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Deleted allocations for instance 7788f9f8-c08e-46f8-b204-070215223436 [ 863.703572] env[61868]: INFO nova.compute.manager [-] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Took 1.02 seconds to deallocate network for instance. [ 863.829386] env[61868]: DEBUG oslo_vmware.api [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Task: {'id': task-1315490, 'name': ReconfigVM_Task, 'duration_secs': 0.398859} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.829688] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Reconfigured VM instance instance-00000044 to attach disk [datastore2] c6f99cbc-3e81-44b4-b3d1-d8cd821829f2/c6f99cbc-3e81-44b4-b3d1-d8cd821829f2.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 863.830320] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bd29bd3a-be1c-4cbb-b3b2-5f4107f26227 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.835838] env[61868]: DEBUG oslo_vmware.api [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Waiting for the task: (returnval){ [ 863.835838] env[61868]: value = "task-1315493" [ 863.835838] env[61868]: _type = "Task" [ 863.835838] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.845815] env[61868]: DEBUG oslo_vmware.api [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Task: {'id': task-1315493, 'name': Rename_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.848657] env[61868]: DEBUG oslo_vmware.api [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Task: {'id': task-1315491, 'name': PowerOffVM_Task, 'duration_secs': 0.183472} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.848828] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 863.849511] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 863.849750] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0e174966-4411-4ae3-a60c-57aea9d98771 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.855015] env[61868]: DEBUG oslo_vmware.api [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Waiting for the task: (returnval){ [ 863.855015] env[61868]: value = "task-1315494" [ 863.855015] env[61868]: _type = "Task" [ 863.855015] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.864357] env[61868]: DEBUG oslo_vmware.api [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Task: {'id': task-1315494, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.864774] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 863.865008] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1630430d-0be3-4dfb-9a36-e30898188582 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.870172] env[61868]: DEBUG oslo_vmware.api [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 863.870172] env[61868]: value = "task-1315495" [ 863.870172] env[61868]: _type = "Task" [ 863.870172] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.877659] env[61868]: DEBUG oslo_vmware.api [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315495, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.069663] env[61868]: DEBUG oslo_vmware.api [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e00042-f598-98db-1bcb-7fc0826019c2, 'name': SearchDatastore_Task, 'duration_secs': 0.012182} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.070009] env[61868]: DEBUG oslo_concurrency.lockutils [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.070259] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 864.070499] env[61868]: DEBUG oslo_concurrency.lockutils [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.070648] env[61868]: DEBUG oslo_concurrency.lockutils [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.070832] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 864.071113] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8f43446c-cef0-4a76-89df-5502b3628461 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.078593] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 864.078776] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 864.079510] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae261e1b-5f11-4adc-a195-9df31d7f40fc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.084173] env[61868]: DEBUG oslo_vmware.api [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 864.084173] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52638b24-44c9-ab88-3812-174e96b7585a" [ 864.084173] env[61868]: _type = "Task" [ 864.084173] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.091944] env[61868]: DEBUG oslo_vmware.api [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52638b24-44c9-ab88-3812-174e96b7585a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.122299] env[61868]: DEBUG oslo_vmware.api [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Task: {'id': task-1315492, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063289} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.122569] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 864.123352] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a29d2cf-b71b-42ff-8673-6b31d1793e6a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.142643] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] cf063dbc-7b5a-4836-91bf-a0aa33cca6bc/cf063dbc-7b5a-4836-91bf-a0aa33cca6bc.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 864.145857] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6191813f-4313-4ec9-935c-44da8800def9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.168583] env[61868]: DEBUG oslo_concurrency.lockutils [None req-05a0085d-ae8b-4e73-9106-330b108967d5 tempest-MultipleCreateTestJSON-1548325849 tempest-MultipleCreateTestJSON-1548325849-project-member] Lock "7788f9f8-c08e-46f8-b204-070215223436" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.319s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.174332] env[61868]: DEBUG oslo_vmware.api [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315489, 'name': ReconfigVM_Task, 'duration_secs': 0.575662} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.175615] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 75fdbc11-1ef3-442d-bcf4-55069ba9ab62/75fdbc11-1ef3-442d-bcf4-55069ba9ab62.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 864.176312] env[61868]: DEBUG oslo_vmware.api [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Waiting for the task: (returnval){ [ 864.176312] env[61868]: value = "task-1315496" [ 864.176312] env[61868]: _type = "Task" [ 864.176312] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.176503] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7be54604-0524-4a0e-beb1-e437d4805f83 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.187711] env[61868]: DEBUG oslo_vmware.api [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Task: {'id': task-1315496, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.191263] env[61868]: DEBUG oslo_vmware.api [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the task: (returnval){ [ 864.191263] env[61868]: value = "task-1315497" [ 864.191263] env[61868]: _type = "Task" [ 864.191263] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.200634] env[61868]: DEBUG oslo_vmware.api [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315497, 'name': Rename_Task} progress is 10%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.210058] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.348346] env[61868]: DEBUG oslo_vmware.api [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Task: {'id': task-1315493, 'name': Rename_Task, 'duration_secs': 0.148999} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.348608] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 864.348867] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-de071c48-149c-4a2c-9923-2131926e59f8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.359936] env[61868]: DEBUG oslo_vmware.api [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Waiting for the task: (returnval){ [ 864.359936] env[61868]: value = "task-1315498" [ 864.359936] env[61868]: _type = "Task" [ 864.359936] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.370377] env[61868]: DEBUG oslo_vmware.api [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Task: {'id': task-1315498, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.378202] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] VM already powered off {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 864.378449] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Volume detach. Driver type: vmdk {{(pid=61868) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 864.378644] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281494', 'volume_id': 'd90113fc-08af-4a33-a1e2-64d3d4549274', 'name': 'volume-d90113fc-08af-4a33-a1e2-64d3d4549274', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2c00210e-3b6a-4041-9e8e-7f40ef23d716', 'attached_at': '', 'detached_at': '', 'volume_id': 'd90113fc-08af-4a33-a1e2-64d3d4549274', 'serial': 'd90113fc-08af-4a33-a1e2-64d3d4549274'} {{(pid=61868) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 864.379718] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0085998-687e-473d-8db9-c320874f7c96 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.390181] env[61868]: DEBUG oslo_vmware.api [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315495, 'name': PowerOffVM_Task, 'duration_secs': 0.217894} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.403917] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 864.407413] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f705e4ca-6f1e-4fe7-a45a-1805f342d192 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.410613] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0557b782-fea2-4d84-93e6-e873786ba2da {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.430671] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1495cb3-65aa-42c3-8020-44c42f372250 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.436323] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b6a69f0-caa1-49e2-8cd7-ee7a0da8f7a9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.463755] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09de8166-b7bb-4fc7-81f1-6dedd3f96baa {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.479458] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] The volume has not been displaced from its original location: [datastore1] volume-d90113fc-08af-4a33-a1e2-64d3d4549274/volume-d90113fc-08af-4a33-a1e2-64d3d4549274.vmdk. No consolidation needed. {{(pid=61868) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 864.484712] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Reconfiguring VM instance instance-00000036 to detach disk 2000 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 864.489283] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 864.489567] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b954e2d6-3f33-438c-b3aa-da1726a184a6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.502009] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-738cb541-4a89-4c27-949a-d03b31bd565b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.508871] env[61868]: DEBUG oslo_vmware.api [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 864.508871] env[61868]: value = "task-1315499" [ 864.508871] env[61868]: _type = "Task" [ 864.508871] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.510315] env[61868]: DEBUG oslo_vmware.api [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Waiting for the task: (returnval){ [ 864.510315] env[61868]: value = "task-1315500" [ 864.510315] env[61868]: _type = "Task" [ 864.510315] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.517557] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4faa1e0f-c692-4688-a716-7e608cb83e62 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.529992] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a18ef0b2-601b-4556-a7f6-f51d2d8b6366 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.533255] env[61868]: DEBUG oslo_vmware.api [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Task: {'id': task-1315500, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.533686] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] VM already powered off {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 864.533794] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 864.534013] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.534180] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.534298] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 864.534885] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ec8209b0-1941-4839-9176-294f4003b450 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.562324] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e1503f5-6ecb-4f96-bb92-f51e19d11349 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.564600] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 864.564789] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 864.565600] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0679e71d-b8c8-4427-aeb4-ba477d923d0c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.573352] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d08c72e9-4b58-480f-8cdc-93b80f469fce {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.577163] env[61868]: DEBUG oslo_vmware.api [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 864.577163] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]523177bf-fec8-b4c8-5aaf-ec687f1583b4" [ 864.577163] env[61868]: _type = "Task" [ 864.577163] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.589649] env[61868]: DEBUG nova.compute.provider_tree [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 864.597630] env[61868]: DEBUG oslo_vmware.api [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]523177bf-fec8-b4c8-5aaf-ec687f1583b4, 'name': SearchDatastore_Task, 'duration_secs': 0.014688} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.599233] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2e4ea35d-032a-4d20-ad5e-9bbd7f09f883 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.605071] env[61868]: DEBUG oslo_vmware.api [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52638b24-44c9-ab88-3812-174e96b7585a, 'name': SearchDatastore_Task, 'duration_secs': 0.00736} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.606105] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e041b05d-05ef-4acf-932c-05b195c34b14 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.609497] env[61868]: DEBUG oslo_vmware.api [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 864.609497] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52020835-c9d8-e9d3-122a-83156f21b4e4" [ 864.609497] env[61868]: _type = "Task" [ 864.609497] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.617017] env[61868]: DEBUG oslo_vmware.api [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 864.617017] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52b3b860-76f8-ecbd-430c-5216b524b711" [ 864.617017] env[61868]: _type = "Task" [ 864.617017] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.620628] env[61868]: DEBUG oslo_vmware.api [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52020835-c9d8-e9d3-122a-83156f21b4e4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.624868] env[61868]: DEBUG oslo_vmware.api [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52b3b860-76f8-ecbd-430c-5216b524b711, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.687498] env[61868]: DEBUG oslo_vmware.api [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Task: {'id': task-1315496, 'name': ReconfigVM_Task, 'duration_secs': 0.385374} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.687788] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Reconfigured VM instance instance-00000045 to attach disk [datastore1] cf063dbc-7b5a-4836-91bf-a0aa33cca6bc/cf063dbc-7b5a-4836-91bf-a0aa33cca6bc.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 864.688410] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-50710511-146b-47c0-96c2-e85ba003034f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.696230] env[61868]: DEBUG oslo_vmware.api [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Waiting for the task: (returnval){ [ 864.696230] env[61868]: value = "task-1315501" [ 864.696230] env[61868]: _type = "Task" [ 864.696230] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.702604] env[61868]: DEBUG oslo_vmware.api [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315497, 'name': Rename_Task, 'duration_secs': 0.154543} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.703276] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 864.703877] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c1c52c7e-1855-4319-9827-c22ec9ffed44 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.708053] env[61868]: DEBUG oslo_vmware.api [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Task: {'id': task-1315501, 'name': Rename_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.712356] env[61868]: DEBUG oslo_vmware.api [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the task: (returnval){ [ 864.712356] env[61868]: value = "task-1315502" [ 864.712356] env[61868]: _type = "Task" [ 864.712356] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.719936] env[61868]: DEBUG oslo_vmware.api [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315502, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.870416] env[61868]: DEBUG oslo_vmware.api [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Task: {'id': task-1315498, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.023114] env[61868]: DEBUG oslo_vmware.api [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Task: {'id': task-1315500, 'name': ReconfigVM_Task, 'duration_secs': 0.273172} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.023344] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Reconfigured VM instance instance-00000036 to detach disk 2000 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 865.028173] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-895109ad-a519-47ab-bde1-919114adfe94 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.042905] env[61868]: DEBUG oslo_vmware.api [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Waiting for the task: (returnval){ [ 865.042905] env[61868]: value = "task-1315503" [ 865.042905] env[61868]: _type = "Task" [ 865.042905] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.051448] env[61868]: DEBUG oslo_vmware.api [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Task: {'id': task-1315503, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.096035] env[61868]: DEBUG nova.scheduler.client.report [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 865.122318] env[61868]: DEBUG oslo_vmware.api [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52020835-c9d8-e9d3-122a-83156f21b4e4, 'name': SearchDatastore_Task, 'duration_secs': 0.018972} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.123039] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.123322] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 19c2720c-90bc-47f6-999b-6031f893408d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d-rescue.vmdk. {{(pid=61868) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 865.123599] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f871a692-0a2c-4e06-8b04-72bef54051c0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.129893] env[61868]: DEBUG oslo_vmware.api [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52b3b860-76f8-ecbd-430c-5216b524b711, 'name': SearchDatastore_Task, 'duration_secs': 0.014863} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.130487] env[61868]: DEBUG oslo_concurrency.lockutils [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.130774] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] c7b10943-5136-44c8-b4e7-59651220a333/c7b10943-5136-44c8-b4e7-59651220a333.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 865.131049] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-41c17838-56cd-4c91-9f6c-11ff629c941b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.134504] env[61868]: DEBUG oslo_vmware.api [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 865.134504] env[61868]: value = "task-1315504" [ 865.134504] env[61868]: _type = "Task" [ 865.134504] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.139165] env[61868]: DEBUG oslo_vmware.api [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 865.139165] env[61868]: value = "task-1315505" [ 865.139165] env[61868]: _type = "Task" [ 865.139165] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.145298] env[61868]: DEBUG oslo_vmware.api [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315504, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.149794] env[61868]: DEBUG oslo_vmware.api [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315505, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.207167] env[61868]: DEBUG oslo_vmware.api [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Task: {'id': task-1315501, 'name': Rename_Task, 'duration_secs': 0.445355} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.207548] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 865.207853] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d724bdf0-08fe-4b66-91f0-7eb462a3230a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.220236] env[61868]: DEBUG oslo_vmware.api [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Waiting for the task: (returnval){ [ 865.220236] env[61868]: value = "task-1315506" [ 865.220236] env[61868]: _type = "Task" [ 865.220236] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.228053] env[61868]: DEBUG oslo_vmware.api [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315502, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.237896] env[61868]: DEBUG oslo_vmware.api [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Task: {'id': task-1315506, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.374600] env[61868]: DEBUG oslo_vmware.api [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Task: {'id': task-1315498, 'name': PowerOnVM_Task, 'duration_secs': 0.861476} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.375009] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 865.375271] env[61868]: INFO nova.compute.manager [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Took 8.01 seconds to spawn the instance on the hypervisor. [ 865.375465] env[61868]: DEBUG nova.compute.manager [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 865.376523] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73af1f6c-724a-4e3a-b6d1-d441f644c89e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.557246] env[61868]: DEBUG oslo_vmware.api [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Task: {'id': task-1315503, 'name': ReconfigVM_Task, 'duration_secs': 0.127379} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.557580] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281494', 'volume_id': 'd90113fc-08af-4a33-a1e2-64d3d4549274', 'name': 'volume-d90113fc-08af-4a33-a1e2-64d3d4549274', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2c00210e-3b6a-4041-9e8e-7f40ef23d716', 'attached_at': '', 'detached_at': '', 'volume_id': 'd90113fc-08af-4a33-a1e2-64d3d4549274', 'serial': 'd90113fc-08af-4a33-a1e2-64d3d4549274'} {{(pid=61868) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 865.557875] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 865.558782] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26bf4242-5444-4076-8d49-312be285bff4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.567110] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 865.567495] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-175b29ce-fe7a-464a-941d-ef2a96ca465a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.601910] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.967s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.602455] env[61868]: DEBUG nova.compute.manager [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 865.605677] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.492s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.609433] env[61868]: INFO nova.compute.claims [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 865.653658] env[61868]: DEBUG oslo_vmware.api [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315504, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.656813] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 865.657126] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 865.657348] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Deleting the datastore file [datastore1] 2c00210e-3b6a-4041-9e8e-7f40ef23d716 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 865.661379] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-38f3b88e-42af-4419-a952-a2bc9d2c1ce1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.668531] env[61868]: DEBUG oslo_vmware.api [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315505, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.674320] env[61868]: DEBUG oslo_vmware.api [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Waiting for the task: (returnval){ [ 865.674320] env[61868]: value = "task-1315508" [ 865.674320] env[61868]: _type = "Task" [ 865.674320] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.687640] env[61868]: DEBUG oslo_vmware.api [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Task: {'id': task-1315508, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.729577] env[61868]: DEBUG oslo_vmware.api [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315502, 'name': PowerOnVM_Task} progress is 87%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.736344] env[61868]: DEBUG oslo_vmware.api [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Task: {'id': task-1315506, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.900581] env[61868]: INFO nova.compute.manager [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Took 20.33 seconds to build instance. [ 866.113132] env[61868]: DEBUG nova.compute.utils [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 866.117863] env[61868]: DEBUG nova.compute.manager [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 866.118168] env[61868]: DEBUG nova.network.neutron [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 866.149915] env[61868]: DEBUG oslo_vmware.api [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315504, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.637532} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.153421] env[61868]: INFO nova.virt.vmwareapi.ds_util [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 19c2720c-90bc-47f6-999b-6031f893408d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d-rescue.vmdk. [ 866.153728] env[61868]: DEBUG oslo_vmware.api [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315505, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.706398} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.154422] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60484d6d-05d9-42a6-9120-0537f37ddaba {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.156786] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] c7b10943-5136-44c8-b4e7-59651220a333/c7b10943-5136-44c8-b4e7-59651220a333.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 866.157040] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 866.157281] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-842b1336-197a-43c9-a15e-2789fad664de {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.186948] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] 19c2720c-90bc-47f6-999b-6031f893408d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d-rescue.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 866.191708] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-692187d3-ef91-42b7-966e-43f8f0aa58b1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.205538] env[61868]: DEBUG oslo_vmware.api [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 866.205538] env[61868]: value = "task-1315509" [ 866.205538] env[61868]: _type = "Task" [ 866.205538] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.207106] env[61868]: DEBUG nova.policy [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c14b4ff01717495ca97c7f35f91c2995', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd7fc76299baf4a97b57139e5f1caa16e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 866.216611] env[61868]: DEBUG oslo_vmware.api [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Task: {'id': task-1315508, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.226885} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.220890] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 866.221100] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 866.221419] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 866.223384] env[61868]: DEBUG oslo_vmware.api [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 866.223384] env[61868]: value = "task-1315510" [ 866.223384] env[61868]: _type = "Task" [ 866.223384] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.226554] env[61868]: DEBUG oslo_vmware.api [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315509, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.239133] env[61868]: DEBUG oslo_vmware.api [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315502, 'name': PowerOnVM_Task, 'duration_secs': 1.282402} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.240225] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 866.240459] env[61868]: INFO nova.compute.manager [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Took 11.30 seconds to spawn the instance on the hypervisor. [ 866.240673] env[61868]: DEBUG nova.compute.manager [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 866.241597] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8b39d47-0941-49f8-a84e-9d050d355598 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.250539] env[61868]: DEBUG oslo_vmware.api [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315510, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.250973] env[61868]: DEBUG oslo_vmware.api [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Task: {'id': task-1315506, 'name': PowerOnVM_Task} progress is 79%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.343271] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Volume detach. Driver type: vmdk {{(pid=61868) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 866.343617] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-158ceceb-931c-4ff8-8850-b9ec52e6be7f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.358020] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22698608-c7db-4c10-9017-ab2bb4697711 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.385155] env[61868]: ERROR nova.compute.manager [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Failed to detach volume d90113fc-08af-4a33-a1e2-64d3d4549274 from /dev/sda: nova.exception.InstanceNotFound: Instance 2c00210e-3b6a-4041-9e8e-7f40ef23d716 could not be found. [ 866.385155] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Traceback (most recent call last): [ 866.385155] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 866.385155] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] self.driver.rebuild(**kwargs) [ 866.385155] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 866.385155] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] raise NotImplementedError() [ 866.385155] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] NotImplementedError [ 866.385155] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] [ 866.385155] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] During handling of the above exception, another exception occurred: [ 866.385155] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] [ 866.385155] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Traceback (most recent call last): [ 866.385155] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 866.385155] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] self.driver.detach_volume(context, old_connection_info, [ 866.385555] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 559, in detach_volume [ 866.385555] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] return self._volumeops.detach_volume(connection_info, instance) [ 866.385555] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 866.385555] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] self._detach_volume_vmdk(connection_info, instance) [ 866.385555] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 866.385555] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 866.385555] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 866.385555] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] stable_ref.fetch_moref(session) [ 866.385555] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 866.385555] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] raise exception.InstanceNotFound(instance_id=self._uuid) [ 866.385555] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] nova.exception.InstanceNotFound: Instance 2c00210e-3b6a-4041-9e8e-7f40ef23d716 could not be found. [ 866.385555] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] [ 866.404853] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1ecc55cf-7b06-4331-a6a0-c7c4549b3940 tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Lock "c6f99cbc-3e81-44b4-b3d1-d8cd821829f2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.744s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.549148] env[61868]: DEBUG nova.compute.utils [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Build of instance 2c00210e-3b6a-4041-9e8e-7f40ef23d716 aborted: Failed to rebuild volume backed instance. {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 866.552864] env[61868]: ERROR nova.compute.manager [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance 2c00210e-3b6a-4041-9e8e-7f40ef23d716 aborted: Failed to rebuild volume backed instance. [ 866.552864] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Traceback (most recent call last): [ 866.552864] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 866.552864] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] self.driver.rebuild(**kwargs) [ 866.552864] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 866.552864] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] raise NotImplementedError() [ 866.552864] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] NotImplementedError [ 866.552864] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] [ 866.552864] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] During handling of the above exception, another exception occurred: [ 866.552864] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] [ 866.552864] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Traceback (most recent call last): [ 866.552864] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] File "/opt/stack/nova/nova/compute/manager.py", line 3600, in _rebuild_volume_backed_instance [ 866.552864] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] self._detach_root_volume(context, instance, root_bdm) [ 866.553495] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] File "/opt/stack/nova/nova/compute/manager.py", line 3579, in _detach_root_volume [ 866.553495] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] with excutils.save_and_reraise_exception(): [ 866.553495] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 866.553495] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] self.force_reraise() [ 866.553495] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 866.553495] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] raise self.value [ 866.553495] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 866.553495] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] self.driver.detach_volume(context, old_connection_info, [ 866.553495] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 559, in detach_volume [ 866.553495] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] return self._volumeops.detach_volume(connection_info, instance) [ 866.553495] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 866.553495] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] self._detach_volume_vmdk(connection_info, instance) [ 866.554193] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 866.554193] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 866.554193] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 866.554193] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] stable_ref.fetch_moref(session) [ 866.554193] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 866.554193] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] raise exception.InstanceNotFound(instance_id=self._uuid) [ 866.554193] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] nova.exception.InstanceNotFound: Instance 2c00210e-3b6a-4041-9e8e-7f40ef23d716 could not be found. [ 866.554193] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] [ 866.554193] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] During handling of the above exception, another exception occurred: [ 866.554193] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] [ 866.554193] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Traceback (most recent call last): [ 866.554193] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] File "/opt/stack/nova/nova/compute/manager.py", line 10865, in _error_out_instance_on_exception [ 866.554193] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] yield [ 866.554193] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] File "/opt/stack/nova/nova/compute/manager.py", line 3868, in rebuild_instance [ 866.555036] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] self._do_rebuild_instance_with_claim( [ 866.555036] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] File "/opt/stack/nova/nova/compute/manager.py", line 3954, in _do_rebuild_instance_with_claim [ 866.555036] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] self._do_rebuild_instance( [ 866.555036] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] File "/opt/stack/nova/nova/compute/manager.py", line 4146, in _do_rebuild_instance [ 866.555036] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] self._rebuild_default_impl(**kwargs) [ 866.555036] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] File "/opt/stack/nova/nova/compute/manager.py", line 3723, in _rebuild_default_impl [ 866.555036] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] self._rebuild_volume_backed_instance( [ 866.555036] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] File "/opt/stack/nova/nova/compute/manager.py", line 3615, in _rebuild_volume_backed_instance [ 866.555036] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] raise exception.BuildAbortException( [ 866.555036] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] nova.exception.BuildAbortException: Build of instance 2c00210e-3b6a-4041-9e8e-7f40ef23d716 aborted: Failed to rebuild volume backed instance. [ 866.555036] env[61868]: ERROR nova.compute.manager [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] [ 866.622291] env[61868]: DEBUG nova.compute.manager [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 866.724794] env[61868]: DEBUG oslo_vmware.api [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315509, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.139855} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.726778] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 866.728182] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-042d6639-ad2a-4bf0-a100-a43338b33410 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.741405] env[61868]: DEBUG oslo_vmware.api [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315510, 'name': ReconfigVM_Task, 'duration_secs': 0.400159} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.756635] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Reconfigured VM instance instance-00000042 to attach disk [datastore2] 19c2720c-90bc-47f6-999b-6031f893408d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d-rescue.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 866.765682] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Reconfiguring VM instance instance-00000040 to attach disk [datastore1] c7b10943-5136-44c8-b4e7-59651220a333/c7b10943-5136-44c8-b4e7-59651220a333.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 866.765808] env[61868]: DEBUG oslo_vmware.api [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Task: {'id': task-1315506, 'name': PowerOnVM_Task, 'duration_secs': 1.437382} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.766737] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14380757-a37d-400a-895f-4439122efb2f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.774082] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d3c9a637-7808-4686-b978-26e6210eaf55 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.791361] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 866.791628] env[61868]: INFO nova.compute.manager [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Took 6.99 seconds to spawn the instance on the hypervisor. [ 866.792591] env[61868]: DEBUG nova.compute.manager [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 866.793179] env[61868]: DEBUG nova.network.neutron [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Successfully created port: 9771cedf-dae3-41e8-ad4b-aec421274d6c {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 866.802816] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a62339b-cdfb-49af-a4e8-30f14f935666 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.807461] env[61868]: INFO nova.compute.manager [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Took 23.30 seconds to build instance. [ 866.854130] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b4597ae2-15c7-45a4-98be-153ae9527187 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.876498] env[61868]: DEBUG oslo_vmware.api [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 866.876498] env[61868]: value = "task-1315511" [ 866.876498] env[61868]: _type = "Task" [ 866.876498] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.885607] env[61868]: DEBUG oslo_vmware.api [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 866.885607] env[61868]: value = "task-1315512" [ 866.885607] env[61868]: _type = "Task" [ 866.885607] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.901087] env[61868]: DEBUG oslo_vmware.api [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315511, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.911072] env[61868]: DEBUG oslo_vmware.api [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315512, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.046062] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-848822ae-ebae-47e8-96ec-2bd4c0979c77 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.055173] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e6002a6-c2a1-4fac-9528-968b2909b579 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.092562] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74ce3442-f71e-47e2-a792-0796bbf8f80c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.101496] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecdc4751-cfc1-4f75-85a0-1700b19f4269 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.117635] env[61868]: DEBUG nova.compute.provider_tree [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 867.311599] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c0183760-b109-42d9-ba8d-31e4ddb8e7dc tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lock "75fdbc11-1ef3-442d-bcf4-55069ba9ab62" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 106.706s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.398619] env[61868]: INFO nova.compute.manager [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Took 20.31 seconds to build instance. [ 867.404642] env[61868]: DEBUG oslo_vmware.api [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315511, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.408292] env[61868]: DEBUG oslo_vmware.api [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315512, 'name': ReconfigVM_Task, 'duration_secs': 0.459158} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.409441] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 867.411317] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8e234415-d4b8-47ec-a9d7-734c9e9b1325 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.418671] env[61868]: DEBUG oslo_vmware.api [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 867.418671] env[61868]: value = "task-1315513" [ 867.418671] env[61868]: _type = "Task" [ 867.418671] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.429643] env[61868]: DEBUG oslo_vmware.api [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315513, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.626028] env[61868]: DEBUG nova.scheduler.client.report [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 867.639023] env[61868]: DEBUG nova.compute.manager [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 867.678144] env[61868]: DEBUG nova.virt.hardware [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 867.678144] env[61868]: DEBUG nova.virt.hardware [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 867.678144] env[61868]: DEBUG nova.virt.hardware [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 867.678332] env[61868]: DEBUG nova.virt.hardware [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 867.678332] env[61868]: DEBUG nova.virt.hardware [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 867.678332] env[61868]: DEBUG nova.virt.hardware [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 867.678332] env[61868]: DEBUG nova.virt.hardware [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 867.682019] env[61868]: DEBUG nova.virt.hardware [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 867.682019] env[61868]: DEBUG nova.virt.hardware [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 867.682019] env[61868]: DEBUG nova.virt.hardware [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 867.682019] env[61868]: DEBUG nova.virt.hardware [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 867.682019] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-899588da-b477-4032-ae43-d9674fa96adc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.691589] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c707db35-331c-41fc-8f8f-142006f0383a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.895834] env[61868]: DEBUG oslo_vmware.api [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315511, 'name': ReconfigVM_Task, 'duration_secs': 0.820794} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.895834] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Reconfigured VM instance instance-00000040 to attach disk [datastore1] c7b10943-5136-44c8-b4e7-59651220a333/c7b10943-5136-44c8-b4e7-59651220a333.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 867.899018] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-42f3c570-f31c-4141-99ef-2720d8e675a9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.907349] env[61868]: DEBUG oslo_concurrency.lockutils [None req-519ee60c-8b38-4676-94dc-ab6193ab5493 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Lock "cf063dbc-7b5a-4836-91bf-a0aa33cca6bc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 77.119s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.907349] env[61868]: DEBUG oslo_vmware.api [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 867.907349] env[61868]: value = "task-1315514" [ 867.907349] env[61868]: _type = "Task" [ 867.907349] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.923229] env[61868]: DEBUG oslo_vmware.api [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315514, 'name': Rename_Task} progress is 10%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.933957] env[61868]: DEBUG oslo_vmware.api [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315513, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.134103] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.528s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.136096] env[61868]: DEBUG nova.compute.manager [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 868.138087] env[61868]: DEBUG oslo_concurrency.lockutils [None req-241ac167-bf6d-4de7-a078-a1e9812e3546 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.598s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.138479] env[61868]: DEBUG nova.objects.instance [None req-241ac167-bf6d-4de7-a078-a1e9812e3546 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Lazy-loading 'resources' on Instance uuid 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 868.420540] env[61868]: DEBUG oslo_vmware.api [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315514, 'name': Rename_Task, 'duration_secs': 0.148719} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.421084] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 868.421482] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e4f6f7af-fba5-458d-8a15-1c3df18003cf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.439187] env[61868]: DEBUG oslo_vmware.api [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315513, 'name': PowerOnVM_Task, 'duration_secs': 0.675903} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.441538] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 868.445514] env[61868]: DEBUG oslo_vmware.api [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 868.445514] env[61868]: value = "task-1315515" [ 868.445514] env[61868]: _type = "Task" [ 868.445514] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.447376] env[61868]: DEBUG nova.compute.manager [None req-e2b0aab3-a38f-4dd8-8ed7-89892e5af932 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 868.449763] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b2bbe31-2264-42cb-b779-7308bedcb10c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.478249] env[61868]: DEBUG oslo_vmware.api [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315515, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.565508] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a1b309cb-b151-464d-8aab-50c6857b088e tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Acquiring lock "c6f99cbc-3e81-44b4-b3d1-d8cd821829f2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.566060] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a1b309cb-b151-464d-8aab-50c6857b088e tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Lock "c6f99cbc-3e81-44b4-b3d1-d8cd821829f2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.566394] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a1b309cb-b151-464d-8aab-50c6857b088e tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Acquiring lock "c6f99cbc-3e81-44b4-b3d1-d8cd821829f2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.566675] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a1b309cb-b151-464d-8aab-50c6857b088e tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Lock "c6f99cbc-3e81-44b4-b3d1-d8cd821829f2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.566959] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a1b309cb-b151-464d-8aab-50c6857b088e tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Lock "c6f99cbc-3e81-44b4-b3d1-d8cd821829f2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.569332] env[61868]: INFO nova.compute.manager [None req-a1b309cb-b151-464d-8aab-50c6857b088e tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Terminating instance [ 868.572516] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.573799] env[61868]: DEBUG nova.compute.manager [None req-a1b309cb-b151-464d-8aab-50c6857b088e tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 868.574178] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a1b309cb-b151-464d-8aab-50c6857b088e tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 868.575800] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc33c06e-c0de-4307-bf39-6ebe29a11bd3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.585695] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1b309cb-b151-464d-8aab-50c6857b088e tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 868.586331] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-049aa87a-21d8-4c54-bd16-de2bbe0db7bf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.593906] env[61868]: DEBUG oslo_vmware.api [None req-a1b309cb-b151-464d-8aab-50c6857b088e tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Waiting for the task: (returnval){ [ 868.593906] env[61868]: value = "task-1315516" [ 868.593906] env[61868]: _type = "Task" [ 868.593906] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.604075] env[61868]: DEBUG oslo_vmware.api [None req-a1b309cb-b151-464d-8aab-50c6857b088e tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Task: {'id': task-1315516, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.642827] env[61868]: DEBUG nova.compute.utils [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 868.651797] env[61868]: DEBUG nova.compute.manager [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 868.652520] env[61868]: DEBUG nova.network.neutron [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 868.701130] env[61868]: DEBUG nova.policy [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7f2eec1c3b204775ae5ffb09518c79e5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '50658d903dee454eb544ebf92621faa2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 868.965043] env[61868]: DEBUG oslo_vmware.api [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315515, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.998171] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquiring lock "e2832c0f-dc62-4628-87d4-e2a01819b771" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.999526] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lock "e2832c0f-dc62-4628-87d4-e2a01819b771" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.054540] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d82b3d5a-f56a-48b2-9dec-c0cf155856b1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.062835] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4587efeb-b0c1-42fd-a86c-9f2386bfc0a3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.098299] env[61868]: DEBUG nova.network.neutron [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Successfully created port: 981ca2e3-c2d7-4d28-bda6-68ddd6d57bc5 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 869.104390] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cba3b335-75ba-4472-bfce-4d64aa38495b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.116353] env[61868]: DEBUG oslo_vmware.api [None req-a1b309cb-b151-464d-8aab-50c6857b088e tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Task: {'id': task-1315516, 'name': PowerOffVM_Task, 'duration_secs': 0.331348} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.119403] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1b309cb-b151-464d-8aab-50c6857b088e tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 869.119403] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a1b309cb-b151-464d-8aab-50c6857b088e tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 869.119403] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5fd3db18-f437-42e4-a6a8-bfe1bcc1ea26 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.121378] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a92a0d7c-0e7f-4c56-8c72-8c2b8b265610 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.138028] env[61868]: DEBUG nova.compute.provider_tree [None req-241ac167-bf6d-4de7-a078-a1e9812e3546 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 869.148423] env[61868]: DEBUG nova.compute.manager [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 869.203362] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a1b309cb-b151-464d-8aab-50c6857b088e tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 869.203362] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a1b309cb-b151-464d-8aab-50c6857b088e tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Deleting contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 869.203362] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1b309cb-b151-464d-8aab-50c6857b088e tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Deleting the datastore file [datastore2] c6f99cbc-3e81-44b4-b3d1-d8cd821829f2 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 869.203362] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6ed59f62-5648-4262-9894-6b9de42290e3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.210098] env[61868]: DEBUG oslo_vmware.api [None req-a1b309cb-b151-464d-8aab-50c6857b088e tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Waiting for the task: (returnval){ [ 869.210098] env[61868]: value = "task-1315518" [ 869.210098] env[61868]: _type = "Task" [ 869.210098] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.218727] env[61868]: DEBUG oslo_vmware.api [None req-a1b309cb-b151-464d-8aab-50c6857b088e tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Task: {'id': task-1315518, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.452601] env[61868]: DEBUG nova.compute.manager [None req-2da6a645-0366-4469-b2d6-08c382b61151 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 869.456773] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5a2342f-0dbe-49ff-bc00-b642f24bfbc5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.468537] env[61868]: DEBUG oslo_vmware.api [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315515, 'name': PowerOnVM_Task, 'duration_secs': 0.620051} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.470850] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 869.471146] env[61868]: DEBUG nova.compute.manager [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 869.472296] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06442daf-e0e6-4c3e-8f4d-ef3a19438101 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.506798] env[61868]: DEBUG nova.compute.manager [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 869.605230] env[61868]: DEBUG nova.network.neutron [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Successfully updated port: 9771cedf-dae3-41e8-ad4b-aec421274d6c {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 869.641131] env[61868]: DEBUG nova.scheduler.client.report [None req-241ac167-bf6d-4de7-a078-a1e9812e3546 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 869.722592] env[61868]: DEBUG oslo_vmware.api [None req-a1b309cb-b151-464d-8aab-50c6857b088e tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Task: {'id': task-1315518, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.201235} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.722909] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1b309cb-b151-464d-8aab-50c6857b088e tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 869.723383] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a1b309cb-b151-464d-8aab-50c6857b088e tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Deleted contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 869.723683] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a1b309cb-b151-464d-8aab-50c6857b088e tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 869.723794] env[61868]: INFO nova.compute.manager [None req-a1b309cb-b151-464d-8aab-50c6857b088e tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Took 1.15 seconds to destroy the instance on the hypervisor. [ 869.724124] env[61868]: DEBUG oslo.service.loopingcall [None req-a1b309cb-b151-464d-8aab-50c6857b088e tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 869.724392] env[61868]: DEBUG nova.compute.manager [-] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 869.724505] env[61868]: DEBUG nova.network.neutron [-] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 869.739247] env[61868]: DEBUG nova.compute.manager [req-30c5cc18-fc03-454e-a9d2-9797c93b89e0 req-3250ebe5-2e27-49dc-8101-e5fc37599dc0 service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Received event network-vif-plugged-9771cedf-dae3-41e8-ad4b-aec421274d6c {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 869.739247] env[61868]: DEBUG oslo_concurrency.lockutils [req-30c5cc18-fc03-454e-a9d2-9797c93b89e0 req-3250ebe5-2e27-49dc-8101-e5fc37599dc0 service nova] Acquiring lock "5aec2609-79d3-4725-a182-70b575adbe17-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.739803] env[61868]: DEBUG oslo_concurrency.lockutils [req-30c5cc18-fc03-454e-a9d2-9797c93b89e0 req-3250ebe5-2e27-49dc-8101-e5fc37599dc0 service nova] Lock "5aec2609-79d3-4725-a182-70b575adbe17-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.739922] env[61868]: DEBUG oslo_concurrency.lockutils [req-30c5cc18-fc03-454e-a9d2-9797c93b89e0 req-3250ebe5-2e27-49dc-8101-e5fc37599dc0 service nova] Lock "5aec2609-79d3-4725-a182-70b575adbe17-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.740184] env[61868]: DEBUG nova.compute.manager [req-30c5cc18-fc03-454e-a9d2-9797c93b89e0 req-3250ebe5-2e27-49dc-8101-e5fc37599dc0 service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] No waiting events found dispatching network-vif-plugged-9771cedf-dae3-41e8-ad4b-aec421274d6c {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 869.740376] env[61868]: WARNING nova.compute.manager [req-30c5cc18-fc03-454e-a9d2-9797c93b89e0 req-3250ebe5-2e27-49dc-8101-e5fc37599dc0 service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Received unexpected event network-vif-plugged-9771cedf-dae3-41e8-ad4b-aec421274d6c for instance with vm_state building and task_state spawning. [ 869.818226] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 869.818479] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 869.972763] env[61868]: INFO nova.compute.manager [None req-2da6a645-0366-4469-b2d6-08c382b61151 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] instance snapshotting [ 869.973967] env[61868]: DEBUG nova.objects.instance [None req-2da6a645-0366-4469-b2d6-08c382b61151 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Lazy-loading 'flavor' on Instance uuid cf063dbc-7b5a-4836-91bf-a0aa33cca6bc {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 869.996561] env[61868]: DEBUG oslo_concurrency.lockutils [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.036233] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.070541] env[61868]: DEBUG nova.compute.manager [req-02ebb4ee-9d81-4a96-bfeb-e123a08920fe req-3d338c9c-f414-4092-baed-eb1054759ccb service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Received event network-changed-9771cedf-dae3-41e8-ad4b-aec421274d6c {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 870.070748] env[61868]: DEBUG nova.compute.manager [req-02ebb4ee-9d81-4a96-bfeb-e123a08920fe req-3d338c9c-f414-4092-baed-eb1054759ccb service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Refreshing instance network info cache due to event network-changed-9771cedf-dae3-41e8-ad4b-aec421274d6c. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 870.070964] env[61868]: DEBUG oslo_concurrency.lockutils [req-02ebb4ee-9d81-4a96-bfeb-e123a08920fe req-3d338c9c-f414-4092-baed-eb1054759ccb service nova] Acquiring lock "refresh_cache-5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.071124] env[61868]: DEBUG oslo_concurrency.lockutils [req-02ebb4ee-9d81-4a96-bfeb-e123a08920fe req-3d338c9c-f414-4092-baed-eb1054759ccb service nova] Acquired lock "refresh_cache-5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.071290] env[61868]: DEBUG nova.network.neutron [req-02ebb4ee-9d81-4a96-bfeb-e123a08920fe req-3d338c9c-f414-4092-baed-eb1054759ccb service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Refreshing network info cache for port 9771cedf-dae3-41e8-ad4b-aec421274d6c {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 870.107070] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "refresh_cache-5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.146226] env[61868]: DEBUG oslo_concurrency.lockutils [None req-241ac167-bf6d-4de7-a078-a1e9812e3546 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.008s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.148510] env[61868]: DEBUG oslo_concurrency.lockutils [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.923s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.150767] env[61868]: INFO nova.compute.claims [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 870.159097] env[61868]: DEBUG nova.compute.manager [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 870.169988] env[61868]: INFO nova.scheduler.client.report [None req-241ac167-bf6d-4de7-a078-a1e9812e3546 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Deleted allocations for instance 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1 [ 870.189975] env[61868]: DEBUG nova.virt.hardware [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 870.190806] env[61868]: DEBUG nova.virt.hardware [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 870.191066] env[61868]: DEBUG nova.virt.hardware [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 870.191324] env[61868]: DEBUG nova.virt.hardware [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 870.191542] env[61868]: DEBUG nova.virt.hardware [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 870.191743] env[61868]: DEBUG nova.virt.hardware [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 870.192267] env[61868]: DEBUG nova.virt.hardware [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 870.192267] env[61868]: DEBUG nova.virt.hardware [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 870.192413] env[61868]: DEBUG nova.virt.hardware [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 870.192619] env[61868]: DEBUG nova.virt.hardware [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 870.192894] env[61868]: DEBUG nova.virt.hardware [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 870.194371] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-306c9f10-75a0-46e8-9a36-f0ce5545b475 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.205384] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32b75277-8db0-4bf6-94d5-62bea6b04b49 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.330462] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 870.332775] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Starting heal instance info cache {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 870.333097] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Rebuilding the list of instances to heal {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 870.381125] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Acquiring lock "cf063dbc-7b5a-4836-91bf-a0aa33cca6bc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.381257] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Lock "cf063dbc-7b5a-4836-91bf-a0aa33cca6bc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.381370] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Acquiring lock "cf063dbc-7b5a-4836-91bf-a0aa33cca6bc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.381540] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Lock "cf063dbc-7b5a-4836-91bf-a0aa33cca6bc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.381806] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Lock "cf063dbc-7b5a-4836-91bf-a0aa33cca6bc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.384194] env[61868]: INFO nova.compute.manager [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Terminating instance [ 870.386053] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Acquiring lock "refresh_cache-cf063dbc-7b5a-4836-91bf-a0aa33cca6bc" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.386226] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Acquired lock "refresh_cache-cf063dbc-7b5a-4836-91bf-a0aa33cca6bc" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.386409] env[61868]: DEBUG nova.network.neutron [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 870.482988] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7d6a01a-4fab-4836-9d3a-8a76bd5eb8d4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.503387] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-845dee14-d8bb-49d2-bc23-1c371444a82f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.608242] env[61868]: DEBUG nova.network.neutron [-] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.640222] env[61868]: DEBUG nova.network.neutron [req-02ebb4ee-9d81-4a96-bfeb-e123a08920fe req-3d338c9c-f414-4092-baed-eb1054759ccb service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 870.682226] env[61868]: DEBUG oslo_concurrency.lockutils [None req-241ac167-bf6d-4de7-a078-a1e9812e3546 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Lock "3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.244s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.745626] env[61868]: DEBUG nova.network.neutron [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Successfully updated port: 981ca2e3-c2d7-4d28-bda6-68ddd6d57bc5 {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 870.836051] env[61868]: DEBUG nova.network.neutron [req-02ebb4ee-9d81-4a96-bfeb-e123a08920fe req-3d338c9c-f414-4092-baed-eb1054759ccb service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.841473] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Skipping network cache update for instance because it is Building. {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 870.841666] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Skipping network cache update for instance because it is being deleted. {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 870.841865] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Skipping network cache update for instance because it is being deleted. {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 870.841994] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Skipping network cache update for instance because it is being deleted. {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 870.843189] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Skipping network cache update for instance because it is Building. {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 870.843355] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Skipping network cache update for instance because it is Building. {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 870.843487] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Skipping network cache update for instance because it is Building. {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 870.909796] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "refresh_cache-2c00210e-3b6a-4041-9e8e-7f40ef23d716" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.911543] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquired lock "refresh_cache-2c00210e-3b6a-4041-9e8e-7f40ef23d716" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.911837] env[61868]: DEBUG nova.network.neutron [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Forcefully refreshing network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 870.912033] env[61868]: DEBUG nova.objects.instance [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lazy-loading 'info_cache' on Instance uuid 2c00210e-3b6a-4041-9e8e-7f40ef23d716 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 870.915288] env[61868]: DEBUG nova.network.neutron [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 870.926513] env[61868]: DEBUG nova.compute.manager [req-1fb33392-6ff1-4fdc-a644-42b7954d1870 req-97c1d1a5-c9e3-4e50-8a38-f3ef916ee9b0 service nova] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Received event network-vif-plugged-981ca2e3-c2d7-4d28-bda6-68ddd6d57bc5 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 870.926513] env[61868]: DEBUG oslo_concurrency.lockutils [req-1fb33392-6ff1-4fdc-a644-42b7954d1870 req-97c1d1a5-c9e3-4e50-8a38-f3ef916ee9b0 service nova] Acquiring lock "c6108cf2-b597-4ca7-8f57-12ea82cc6c2c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.926513] env[61868]: DEBUG oslo_concurrency.lockutils [req-1fb33392-6ff1-4fdc-a644-42b7954d1870 req-97c1d1a5-c9e3-4e50-8a38-f3ef916ee9b0 service nova] Lock "c6108cf2-b597-4ca7-8f57-12ea82cc6c2c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.926513] env[61868]: DEBUG oslo_concurrency.lockutils [req-1fb33392-6ff1-4fdc-a644-42b7954d1870 req-97c1d1a5-c9e3-4e50-8a38-f3ef916ee9b0 service nova] Lock "c6108cf2-b597-4ca7-8f57-12ea82cc6c2c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.926513] env[61868]: DEBUG nova.compute.manager [req-1fb33392-6ff1-4fdc-a644-42b7954d1870 req-97c1d1a5-c9e3-4e50-8a38-f3ef916ee9b0 service nova] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] No waiting events found dispatching network-vif-plugged-981ca2e3-c2d7-4d28-bda6-68ddd6d57bc5 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 870.926734] env[61868]: WARNING nova.compute.manager [req-1fb33392-6ff1-4fdc-a644-42b7954d1870 req-97c1d1a5-c9e3-4e50-8a38-f3ef916ee9b0 service nova] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Received unexpected event network-vif-plugged-981ca2e3-c2d7-4d28-bda6-68ddd6d57bc5 for instance with vm_state building and task_state spawning. [ 870.972828] env[61868]: DEBUG nova.network.neutron [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.015647] env[61868]: DEBUG nova.compute.manager [None req-2da6a645-0366-4469-b2d6-08c382b61151 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Instance disappeared during snapshot {{(pid=61868) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 871.059249] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Acquiring lock "8ea8e28b-3582-45f2-b4b5-84f624415a58" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.059562] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Lock "8ea8e28b-3582-45f2-b4b5-84f624415a58" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.064023] env[61868]: DEBUG oslo_concurrency.lockutils [None req-590c00d1-2bb3-45f8-b6e0-454b3a95e2bc tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Acquiring lock "2c00210e-3b6a-4041-9e8e-7f40ef23d716" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.064023] env[61868]: DEBUG oslo_concurrency.lockutils [None req-590c00d1-2bb3-45f8-b6e0-454b3a95e2bc tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Lock "2c00210e-3b6a-4041-9e8e-7f40ef23d716" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.064023] env[61868]: DEBUG oslo_concurrency.lockutils [None req-590c00d1-2bb3-45f8-b6e0-454b3a95e2bc tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Acquiring lock "2c00210e-3b6a-4041-9e8e-7f40ef23d716-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.064023] env[61868]: DEBUG oslo_concurrency.lockutils [None req-590c00d1-2bb3-45f8-b6e0-454b3a95e2bc tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Lock "2c00210e-3b6a-4041-9e8e-7f40ef23d716-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.064204] env[61868]: DEBUG oslo_concurrency.lockutils [None req-590c00d1-2bb3-45f8-b6e0-454b3a95e2bc tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Lock "2c00210e-3b6a-4041-9e8e-7f40ef23d716-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.064204] env[61868]: INFO nova.compute.manager [None req-590c00d1-2bb3-45f8-b6e0-454b3a95e2bc tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Terminating instance [ 871.066165] env[61868]: DEBUG nova.compute.manager [None req-590c00d1-2bb3-45f8-b6e0-454b3a95e2bc tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 871.066442] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d5cecf60-bc33-4946-92b2-471997818a3b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.081121] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cdf3fba-68bd-4438-9b5a-f823db83db23 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.108904] env[61868]: WARNING nova.virt.vmwareapi.driver [None req-590c00d1-2bb3-45f8-b6e0-454b3a95e2bc tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 2c00210e-3b6a-4041-9e8e-7f40ef23d716 could not be found. [ 871.109980] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-590c00d1-2bb3-45f8-b6e0-454b3a95e2bc tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 871.109980] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-37a41155-1aad-43eb-a77a-7e40142c9c21 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.111820] env[61868]: INFO nova.compute.manager [-] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Took 1.39 seconds to deallocate network for instance. [ 871.120769] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89203f15-ce8f-4f83-9949-9fcf39f1386d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.163731] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-590c00d1-2bb3-45f8-b6e0-454b3a95e2bc tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2c00210e-3b6a-4041-9e8e-7f40ef23d716 could not be found. [ 871.164067] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-590c00d1-2bb3-45f8-b6e0-454b3a95e2bc tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 871.164335] env[61868]: INFO nova.compute.manager [None req-590c00d1-2bb3-45f8-b6e0-454b3a95e2bc tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Took 0.10 seconds to destroy the instance on the hypervisor. [ 871.164691] env[61868]: DEBUG oslo.service.loopingcall [None req-590c00d1-2bb3-45f8-b6e0-454b3a95e2bc tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 871.166244] env[61868]: DEBUG nova.compute.manager [-] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 871.166364] env[61868]: DEBUG nova.network.neutron [-] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 871.174719] env[61868]: DEBUG nova.compute.manager [None req-2da6a645-0366-4469-b2d6-08c382b61151 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Found 0 images (rotation: 2) {{(pid=61868) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 871.250743] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Acquiring lock "refresh_cache-c6108cf2-b597-4ca7-8f57-12ea82cc6c2c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.250890] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Acquired lock "refresh_cache-c6108cf2-b597-4ca7-8f57-12ea82cc6c2c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.251070] env[61868]: DEBUG nova.network.neutron [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 871.346928] env[61868]: DEBUG oslo_concurrency.lockutils [req-02ebb4ee-9d81-4a96-bfeb-e123a08920fe req-3d338c9c-f414-4092-baed-eb1054759ccb service nova] Releasing lock "refresh_cache-5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.347512] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquired lock "refresh_cache-5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.347759] env[61868]: DEBUG nova.network.neutron [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 871.398686] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9d9434da-029e-4450-831a-50e8cce87561 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "c7b10943-5136-44c8-b4e7-59651220a333" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.399049] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9d9434da-029e-4450-831a-50e8cce87561 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "c7b10943-5136-44c8-b4e7-59651220a333" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.399585] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9d9434da-029e-4450-831a-50e8cce87561 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "c7b10943-5136-44c8-b4e7-59651220a333-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.400050] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9d9434da-029e-4450-831a-50e8cce87561 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "c7b10943-5136-44c8-b4e7-59651220a333-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.400283] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9d9434da-029e-4450-831a-50e8cce87561 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "c7b10943-5136-44c8-b4e7-59651220a333-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.402810] env[61868]: INFO nova.compute.manager [None req-9d9434da-029e-4450-831a-50e8cce87561 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Terminating instance [ 871.409989] env[61868]: DEBUG nova.compute.manager [None req-9d9434da-029e-4450-831a-50e8cce87561 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 871.410106] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9d9434da-029e-4450-831a-50e8cce87561 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 871.411086] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee2a1cfd-99fd-4dca-a6c2-8e10440eb71e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.424694] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d9434da-029e-4450-831a-50e8cce87561 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 871.424849] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-50681ec3-bc94-4c8c-8066-8206e03ffc31 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.432385] env[61868]: DEBUG oslo_vmware.api [None req-9d9434da-029e-4450-831a-50e8cce87561 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 871.432385] env[61868]: value = "task-1315519" [ 871.432385] env[61868]: _type = "Task" [ 871.432385] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.441552] env[61868]: DEBUG oslo_vmware.api [None req-9d9434da-029e-4450-831a-50e8cce87561 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315519, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.475689] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Releasing lock "refresh_cache-cf063dbc-7b5a-4836-91bf-a0aa33cca6bc" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.476147] env[61868]: DEBUG nova.compute.manager [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 871.476469] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 871.477556] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22591569-b546-434f-8841-661d1625e83e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.485952] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 871.489771] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2ba17f44-c065-4e0e-9351-5c494dd38eca {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.496732] env[61868]: DEBUG oslo_vmware.api [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Waiting for the task: (returnval){ [ 871.496732] env[61868]: value = "task-1315520" [ 871.496732] env[61868]: _type = "Task" [ 871.496732] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.506565] env[61868]: DEBUG oslo_vmware.api [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Task: {'id': task-1315520, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.545575] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d93c19a-a3f3-43d2-9ebc-e6bf7b563105 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.552716] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c685a1b-9819-468a-ab74-d0ce7e49bab1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.586464] env[61868]: DEBUG nova.compute.manager [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 871.593071] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3f7495f-c13c-4455-8f9f-2e6912a931f5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.598434] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d080cd5-0d69-426d-a63e-cde8d621a0b4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.614213] env[61868]: DEBUG nova.compute.provider_tree [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 871.619413] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a1b309cb-b151-464d-8aab-50c6857b088e tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.897409] env[61868]: DEBUG nova.network.neutron [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 871.948071] env[61868]: DEBUG oslo_vmware.api [None req-9d9434da-029e-4450-831a-50e8cce87561 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315519, 'name': PowerOffVM_Task, 'duration_secs': 0.263734} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.948071] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d9434da-029e-4450-831a-50e8cce87561 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 871.948071] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9d9434da-029e-4450-831a-50e8cce87561 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 871.948071] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-35fd238f-ff73-427a-9b3d-e2776d2c8bb4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.980045] env[61868]: DEBUG nova.network.neutron [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 872.010061] env[61868]: DEBUG oslo_vmware.api [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Task: {'id': task-1315520, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.020041] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9d9434da-029e-4450-831a-50e8cce87561 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 872.020041] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9d9434da-029e-4450-831a-50e8cce87561 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 872.020320] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d9434da-029e-4450-831a-50e8cce87561 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Deleting the datastore file [datastore1] c7b10943-5136-44c8-b4e7-59651220a333 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 872.020516] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1cce7e6e-f09a-4c58-a68f-7efd8f3975b5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.029785] env[61868]: DEBUG oslo_vmware.api [None req-9d9434da-029e-4450-831a-50e8cce87561 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 872.029785] env[61868]: value = "task-1315522" [ 872.029785] env[61868]: _type = "Task" [ 872.029785] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.040264] env[61868]: DEBUG oslo_vmware.api [None req-9d9434da-029e-4450-831a-50e8cce87561 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315522, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.110730] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.122099] env[61868]: DEBUG nova.scheduler.client.report [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 872.128779] env[61868]: DEBUG nova.network.neutron [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Updating instance_info_cache with network_info: [{"id": "981ca2e3-c2d7-4d28-bda6-68ddd6d57bc5", "address": "fa:16:3e:d8:2f:78", "network": {"id": "e3a6a163-518e-43ca-b17d-d38d64a3d2a4", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1496913280-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "50658d903dee454eb544ebf92621faa2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24210a23-d8ac-4f4f-84ac-dc0636de9a72", "external-id": "nsx-vlan-transportzone-257", "segmentation_id": 257, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap981ca2e3-c2", "ovs_interfaceid": "981ca2e3-c2d7-4d28-bda6-68ddd6d57bc5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.210403] env[61868]: DEBUG nova.network.neutron [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Updating instance_info_cache with network_info: [{"id": "9771cedf-dae3-41e8-ad4b-aec421274d6c", "address": "fa:16:3e:4f:5f:b8", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9771cedf-da", "ovs_interfaceid": "9771cedf-dae3-41e8-ad4b-aec421274d6c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.247474] env[61868]: DEBUG nova.network.neutron [-] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.346071] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "842c8606-632a-47d4-9a67-38d8e6ce6b4d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.346329] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "842c8606-632a-47d4-9a67-38d8e6ce6b4d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.356767] env[61868]: DEBUG nova.compute.manager [req-52887e7c-ceeb-492c-8739-7bfaf6811b36 req-534e34f8-3663-4e01-88b9-b9cfd29e5d76 service nova] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Received event network-vif-deleted-5159fed6-c58a-4dc0-9b7d-964aa28faaef {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 872.518483] env[61868]: DEBUG oslo_vmware.api [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Task: {'id': task-1315520, 'name': PowerOffVM_Task, 'duration_secs': 0.969234} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.518483] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 872.518483] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 872.518483] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-972d6a78-8e2e-415d-8b5d-a32f500590c8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.539545] env[61868]: DEBUG oslo_vmware.api [None req-9d9434da-029e-4450-831a-50e8cce87561 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315522, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.612545] env[61868]: DEBUG nova.network.neutron [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.629030] env[61868]: DEBUG oslo_concurrency.lockutils [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.478s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.629030] env[61868]: DEBUG nova.compute.manager [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 872.631069] env[61868]: DEBUG oslo_concurrency.lockutils [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.687s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.631788] env[61868]: INFO nova.compute.claims [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 872.635391] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Releasing lock "refresh_cache-c6108cf2-b597-4ca7-8f57-12ea82cc6c2c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.635391] env[61868]: DEBUG nova.compute.manager [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Instance network_info: |[{"id": "981ca2e3-c2d7-4d28-bda6-68ddd6d57bc5", "address": "fa:16:3e:d8:2f:78", "network": {"id": "e3a6a163-518e-43ca-b17d-d38d64a3d2a4", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1496913280-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "50658d903dee454eb544ebf92621faa2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24210a23-d8ac-4f4f-84ac-dc0636de9a72", "external-id": "nsx-vlan-transportzone-257", "segmentation_id": 257, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap981ca2e3-c2", "ovs_interfaceid": "981ca2e3-c2d7-4d28-bda6-68ddd6d57bc5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 872.636232] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d8:2f:78', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24210a23-d8ac-4f4f-84ac-dc0636de9a72', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '981ca2e3-c2d7-4d28-bda6-68ddd6d57bc5', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 872.644644] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Creating folder: Project (50658d903dee454eb544ebf92621faa2). Parent ref: group-v281478. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 872.645052] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4a3a77c2-87a7-4423-8676-d1a755973682 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.656222] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Created folder: Project (50658d903dee454eb544ebf92621faa2) in parent group-v281478. [ 872.656450] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Creating folder: Instances. Parent ref: group-v281542. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 872.656719] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e6940ec1-4112-4601-863f-c70f9f22160a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.666416] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Created folder: Instances in parent group-v281542. [ 872.666694] env[61868]: DEBUG oslo.service.loopingcall [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 872.666920] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 872.667154] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b2ebb0a2-13b9-4b80-afb9-e501119dea09 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.693798] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 872.693798] env[61868]: value = "task-1315526" [ 872.693798] env[61868]: _type = "Task" [ 872.693798] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.707477] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315526, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.713315] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Releasing lock "refresh_cache-5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.713315] env[61868]: DEBUG nova.compute.manager [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Instance network_info: |[{"id": "9771cedf-dae3-41e8-ad4b-aec421274d6c", "address": "fa:16:3e:4f:5f:b8", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9771cedf-da", "ovs_interfaceid": "9771cedf-dae3-41e8-ad4b-aec421274d6c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 872.714028] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4f:5f:b8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8b29df12-5674-476d-a9e5-5e20f704d224', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9771cedf-dae3-41e8-ad4b-aec421274d6c', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 872.722518] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Creating folder: Project (d7fc76299baf4a97b57139e5f1caa16e). Parent ref: group-v281478. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 872.724156] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c927b36c-a4ee-43ae-8d53-27fbf6762cc2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.726469] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 872.726469] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 872.726732] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Deleting the datastore file [datastore1] cf063dbc-7b5a-4836-91bf-a0aa33cca6bc {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 872.727338] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5fff26fa-b08d-4b0e-b27b-54c164eca4f6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.733764] env[61868]: DEBUG oslo_vmware.api [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Waiting for the task: (returnval){ [ 872.733764] env[61868]: value = "task-1315528" [ 872.733764] env[61868]: _type = "Task" [ 872.733764] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.738306] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Created folder: Project (d7fc76299baf4a97b57139e5f1caa16e) in parent group-v281478. [ 872.738499] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Creating folder: Instances. Parent ref: group-v281545. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 872.739066] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-eb59f7ad-2e02-4883-ba3d-497c576c5b45 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.743360] env[61868]: DEBUG oslo_vmware.api [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Task: {'id': task-1315528, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.750275] env[61868]: INFO nova.compute.manager [-] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Took 1.58 seconds to deallocate network for instance. [ 872.754153] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Created folder: Instances in parent group-v281545. [ 872.754346] env[61868]: DEBUG oslo.service.loopingcall [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 872.754551] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 872.754795] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-72d41d29-df56-4105-b538-33fc2efd2e3d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.775858] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 872.775858] env[61868]: value = "task-1315530" [ 872.775858] env[61868]: _type = "Task" [ 872.775858] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.784822] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315530, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.848443] env[61868]: DEBUG nova.compute.manager [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 873.044337] env[61868]: DEBUG oslo_vmware.api [None req-9d9434da-029e-4450-831a-50e8cce87561 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315522, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.803388} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.044630] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d9434da-029e-4450-831a-50e8cce87561 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 873.044821] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9d9434da-029e-4450-831a-50e8cce87561 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 873.045021] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9d9434da-029e-4450-831a-50e8cce87561 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 873.045487] env[61868]: INFO nova.compute.manager [None req-9d9434da-029e-4450-831a-50e8cce87561 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Took 1.64 seconds to destroy the instance on the hypervisor. [ 873.045603] env[61868]: DEBUG oslo.service.loopingcall [None req-9d9434da-029e-4450-831a-50e8cce87561 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 873.045805] env[61868]: DEBUG nova.compute.manager [-] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 873.045911] env[61868]: DEBUG nova.network.neutron [-] [instance: c7b10943-5136-44c8-b4e7-59651220a333] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 873.049927] env[61868]: DEBUG nova.compute.manager [req-4253dabb-1003-43aa-bec2-ee2e72cc52f0 req-6ee79761-8027-46a6-9d6e-ae8cbfc8b79a service nova] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Received event network-changed-981ca2e3-c2d7-4d28-bda6-68ddd6d57bc5 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 873.050200] env[61868]: DEBUG nova.compute.manager [req-4253dabb-1003-43aa-bec2-ee2e72cc52f0 req-6ee79761-8027-46a6-9d6e-ae8cbfc8b79a service nova] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Refreshing instance network info cache due to event network-changed-981ca2e3-c2d7-4d28-bda6-68ddd6d57bc5. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 873.050477] env[61868]: DEBUG oslo_concurrency.lockutils [req-4253dabb-1003-43aa-bec2-ee2e72cc52f0 req-6ee79761-8027-46a6-9d6e-ae8cbfc8b79a service nova] Acquiring lock "refresh_cache-c6108cf2-b597-4ca7-8f57-12ea82cc6c2c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.050568] env[61868]: DEBUG oslo_concurrency.lockutils [req-4253dabb-1003-43aa-bec2-ee2e72cc52f0 req-6ee79761-8027-46a6-9d6e-ae8cbfc8b79a service nova] Acquired lock "refresh_cache-c6108cf2-b597-4ca7-8f57-12ea82cc6c2c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.050725] env[61868]: DEBUG nova.network.neutron [req-4253dabb-1003-43aa-bec2-ee2e72cc52f0 req-6ee79761-8027-46a6-9d6e-ae8cbfc8b79a service nova] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Refreshing network info cache for port 981ca2e3-c2d7-4d28-bda6-68ddd6d57bc5 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 873.114111] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Releasing lock "refresh_cache-2c00210e-3b6a-4041-9e8e-7f40ef23d716" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.114354] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Updated the network info_cache for instance {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 873.114566] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 873.114722] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 873.115894] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 873.115894] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 873.115894] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 873.115894] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 873.115894] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61868) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 873.115894] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager.update_available_resource {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 873.136910] env[61868]: DEBUG nova.compute.utils [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 873.143145] env[61868]: DEBUG nova.compute.manager [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 873.143145] env[61868]: DEBUG nova.network.neutron [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 873.207611] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315526, 'name': CreateVM_Task, 'duration_secs': 0.308861} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.207797] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 873.208663] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.208833] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.209183] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 873.209440] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-189324db-1fc1-431b-b364-8fad53d7aca0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.213198] env[61868]: DEBUG nova.policy [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f73bfcdddbd44790aee8d521241f1991', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e313c27c7d7342e3bc6501653c0c685d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 873.216202] env[61868]: DEBUG oslo_vmware.api [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Waiting for the task: (returnval){ [ 873.216202] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5280c7c3-3d57-c7a8-06c1-76cad7c93897" [ 873.216202] env[61868]: _type = "Task" [ 873.216202] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.226035] env[61868]: DEBUG oslo_vmware.api [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5280c7c3-3d57-c7a8-06c1-76cad7c93897, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.242631] env[61868]: DEBUG oslo_vmware.api [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Task: {'id': task-1315528, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.112384} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.242941] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 873.243147] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 873.243323] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 873.243497] env[61868]: INFO nova.compute.manager [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Took 1.77 seconds to destroy the instance on the hypervisor. [ 873.243752] env[61868]: DEBUG oslo.service.loopingcall [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 873.243955] env[61868]: DEBUG nova.compute.manager [-] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 873.244062] env[61868]: DEBUG nova.network.neutron [-] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 873.272255] env[61868]: DEBUG nova.network.neutron [-] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 873.273071] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "e59a182d-97b2-454a-bc40-8afb0839324f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.273299] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "e59a182d-97b2-454a-bc40-8afb0839324f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.288821] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315530, 'name': CreateVM_Task, 'duration_secs': 0.304754} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.289091] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 873.290277] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.327398] env[61868]: INFO nova.compute.manager [None req-590c00d1-2bb3-45f8-b6e0-454b3a95e2bc tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Took 0.58 seconds to detach 1 volumes for instance. [ 873.331334] env[61868]: DEBUG nova.compute.manager [None req-590c00d1-2bb3-45f8-b6e0-454b3a95e2bc tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Deleting volume: d90113fc-08af-4a33-a1e2-64d3d4549274 {{(pid=61868) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 873.389506] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.622420] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.642403] env[61868]: DEBUG nova.compute.manager [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 873.732692] env[61868]: DEBUG oslo_vmware.api [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5280c7c3-3d57-c7a8-06c1-76cad7c93897, 'name': SearchDatastore_Task, 'duration_secs': 0.017438} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.733603] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.734157] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 873.734157] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.734467] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.734785] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 873.735051] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.735445] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 873.736380] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-76f39fbb-72a1-46b5-a023-227234ad42c9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.739394] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-53db6420-b213-4f0c-b5f5-1016d6253a1b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.748419] env[61868]: DEBUG oslo_vmware.api [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for the task: (returnval){ [ 873.748419] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e0cc4c-04c7-4266-9aca-0d50e5a5c898" [ 873.748419] env[61868]: _type = "Task" [ 873.748419] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.750402] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 873.750642] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 873.754285] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d5f2ae3-7cf8-403c-8b2e-44075e9a9ea2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.758179] env[61868]: DEBUG nova.network.neutron [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Successfully created port: ae9b8f9e-17ec-4c1f-8052-8467a711410c {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 873.769967] env[61868]: DEBUG oslo_vmware.api [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Waiting for the task: (returnval){ [ 873.769967] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]529eab37-05b1-d0c1-247a-628d08b99d8f" [ 873.769967] env[61868]: _type = "Task" [ 873.769967] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.770403] env[61868]: DEBUG oslo_vmware.api [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e0cc4c-04c7-4266-9aca-0d50e5a5c898, 'name': SearchDatastore_Task, 'duration_secs': 0.013759} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.771128] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.771128] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 873.771305] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.775625] env[61868]: DEBUG nova.network.neutron [-] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.781228] env[61868]: DEBUG oslo_vmware.api [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]529eab37-05b1-d0c1-247a-628d08b99d8f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.800345] env[61868]: DEBUG nova.network.neutron [-] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.896661] env[61868]: DEBUG oslo_concurrency.lockutils [None req-590c00d1-2bb3-45f8-b6e0-454b3a95e2bc tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.048257] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f88f609-6415-4240-a49e-8070f6be97a3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.059232] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b404d5c-dc8d-4cb4-81cc-92071136713c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.089784] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9102dafb-a3d4-4999-bf98-1d21bd93531a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.099214] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-196d9e71-69ec-44ae-b634-2a74398d733b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.115022] env[61868]: DEBUG nova.compute.provider_tree [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 874.131401] env[61868]: DEBUG nova.network.neutron [req-4253dabb-1003-43aa-bec2-ee2e72cc52f0 req-6ee79761-8027-46a6-9d6e-ae8cbfc8b79a service nova] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Updated VIF entry in instance network info cache for port 981ca2e3-c2d7-4d28-bda6-68ddd6d57bc5. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 874.131839] env[61868]: DEBUG nova.network.neutron [req-4253dabb-1003-43aa-bec2-ee2e72cc52f0 req-6ee79761-8027-46a6-9d6e-ae8cbfc8b79a service nova] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Updating instance_info_cache with network_info: [{"id": "981ca2e3-c2d7-4d28-bda6-68ddd6d57bc5", "address": "fa:16:3e:d8:2f:78", "network": {"id": "e3a6a163-518e-43ca-b17d-d38d64a3d2a4", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1496913280-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "50658d903dee454eb544ebf92621faa2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24210a23-d8ac-4f4f-84ac-dc0636de9a72", "external-id": "nsx-vlan-transportzone-257", "segmentation_id": 257, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap981ca2e3-c2", "ovs_interfaceid": "981ca2e3-c2d7-4d28-bda6-68ddd6d57bc5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.199728] env[61868]: DEBUG oslo_concurrency.lockutils [None req-62918748-546f-4030-9046-6bb7ec988ab2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Acquiring lock "52a5afc6-4e79-436c-bc94-b61ca9fb860c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.200639] env[61868]: DEBUG oslo_concurrency.lockutils [None req-62918748-546f-4030-9046-6bb7ec988ab2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Lock "52a5afc6-4e79-436c-bc94-b61ca9fb860c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.200839] env[61868]: DEBUG oslo_concurrency.lockutils [None req-62918748-546f-4030-9046-6bb7ec988ab2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Acquiring lock "52a5afc6-4e79-436c-bc94-b61ca9fb860c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.201049] env[61868]: DEBUG oslo_concurrency.lockutils [None req-62918748-546f-4030-9046-6bb7ec988ab2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Lock "52a5afc6-4e79-436c-bc94-b61ca9fb860c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.201232] env[61868]: DEBUG oslo_concurrency.lockutils [None req-62918748-546f-4030-9046-6bb7ec988ab2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Lock "52a5afc6-4e79-436c-bc94-b61ca9fb860c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.203447] env[61868]: INFO nova.compute.manager [None req-62918748-546f-4030-9046-6bb7ec988ab2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Terminating instance [ 874.205771] env[61868]: DEBUG nova.compute.manager [None req-62918748-546f-4030-9046-6bb7ec988ab2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 874.205976] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-62918748-546f-4030-9046-6bb7ec988ab2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 874.206826] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-072e8e5f-137d-4310-affa-56324597f7e4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.215716] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-62918748-546f-4030-9046-6bb7ec988ab2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 874.216070] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e4e12789-c181-4b7f-bbb4-37c00b0e0180 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.223046] env[61868]: DEBUG oslo_vmware.api [None req-62918748-546f-4030-9046-6bb7ec988ab2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Waiting for the task: (returnval){ [ 874.223046] env[61868]: value = "task-1315532" [ 874.223046] env[61868]: _type = "Task" [ 874.223046] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.235381] env[61868]: DEBUG oslo_vmware.api [None req-62918748-546f-4030-9046-6bb7ec988ab2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315532, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.285081] env[61868]: INFO nova.compute.manager [-] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Took 1.04 seconds to deallocate network for instance. [ 874.286639] env[61868]: DEBUG oslo_vmware.api [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]529eab37-05b1-d0c1-247a-628d08b99d8f, 'name': SearchDatastore_Task, 'duration_secs': 0.024549} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.292919] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d307bca7-16d5-4f9d-8ac7-9aa6eac25303 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.299682] env[61868]: DEBUG oslo_vmware.api [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Waiting for the task: (returnval){ [ 874.299682] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52464826-7b2d-57dd-4630-0d60519116f0" [ 874.299682] env[61868]: _type = "Task" [ 874.299682] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.305779] env[61868]: INFO nova.compute.manager [-] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Took 1.26 seconds to deallocate network for instance. [ 874.316267] env[61868]: DEBUG oslo_vmware.api [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52464826-7b2d-57dd-4630-0d60519116f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.473694] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a76b1c97-d8a9-4316-9001-c2415aa6d377 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Acquiring lock "c2693a43-4ea2-4ab0-8915-2fa544780e3c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.474031] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a76b1c97-d8a9-4316-9001-c2415aa6d377 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Lock "c2693a43-4ea2-4ab0-8915-2fa544780e3c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.474268] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a76b1c97-d8a9-4316-9001-c2415aa6d377 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Acquiring lock "c2693a43-4ea2-4ab0-8915-2fa544780e3c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.474468] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a76b1c97-d8a9-4316-9001-c2415aa6d377 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Lock "c2693a43-4ea2-4ab0-8915-2fa544780e3c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.474646] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a76b1c97-d8a9-4316-9001-c2415aa6d377 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Lock "c2693a43-4ea2-4ab0-8915-2fa544780e3c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.476821] env[61868]: INFO nova.compute.manager [None req-a76b1c97-d8a9-4316-9001-c2415aa6d377 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Terminating instance [ 874.478657] env[61868]: DEBUG nova.compute.manager [None req-a76b1c97-d8a9-4316-9001-c2415aa6d377 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 874.478918] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a76b1c97-d8a9-4316-9001-c2415aa6d377 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 874.480124] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e8041ac-9623-47a5-b7a5-82cc58d3705a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.490440] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-a76b1c97-d8a9-4316-9001-c2415aa6d377 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 874.490799] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4fe24b52-cbe9-48d9-85af-d6f279060e5c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.498545] env[61868]: DEBUG oslo_vmware.api [None req-a76b1c97-d8a9-4316-9001-c2415aa6d377 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Waiting for the task: (returnval){ [ 874.498545] env[61868]: value = "task-1315533" [ 874.498545] env[61868]: _type = "Task" [ 874.498545] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.506881] env[61868]: DEBUG oslo_vmware.api [None req-a76b1c97-d8a9-4316-9001-c2415aa6d377 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315533, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.540838] env[61868]: DEBUG nova.compute.manager [req-9486f9ef-cb8b-487c-a586-f9e5248719a9 req-2673899a-55dd-4a88-9445-216236e4399d service nova] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Received event network-vif-deleted-7b94f787-9d7e-4edf-92b0-44d37a7e93ec {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 874.616189] env[61868]: DEBUG nova.scheduler.client.report [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 874.634703] env[61868]: DEBUG oslo_concurrency.lockutils [req-4253dabb-1003-43aa-bec2-ee2e72cc52f0 req-6ee79761-8027-46a6-9d6e-ae8cbfc8b79a service nova] Releasing lock "refresh_cache-c6108cf2-b597-4ca7-8f57-12ea82cc6c2c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.635030] env[61868]: DEBUG nova.compute.manager [req-4253dabb-1003-43aa-bec2-ee2e72cc52f0 req-6ee79761-8027-46a6-9d6e-ae8cbfc8b79a service nova] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Received event network-vif-deleted-bf06b712-731e-430b-bec0-8f026c825d33 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 874.671186] env[61868]: DEBUG nova.compute.manager [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 874.698459] env[61868]: DEBUG nova.virt.hardware [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 874.698628] env[61868]: DEBUG nova.virt.hardware [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 874.698817] env[61868]: DEBUG nova.virt.hardware [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 874.698996] env[61868]: DEBUG nova.virt.hardware [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 874.699183] env[61868]: DEBUG nova.virt.hardware [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 874.699355] env[61868]: DEBUG nova.virt.hardware [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 874.699581] env[61868]: DEBUG nova.virt.hardware [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 874.699765] env[61868]: DEBUG nova.virt.hardware [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 874.699960] env[61868]: DEBUG nova.virt.hardware [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 874.700271] env[61868]: DEBUG nova.virt.hardware [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 874.700495] env[61868]: DEBUG nova.virt.hardware [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 874.701399] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3154b3e-855e-4548-9669-477d25f3ec57 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.710329] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ddead3c-99c7-4492-9f4f-22595cd59610 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.732481] env[61868]: DEBUG oslo_vmware.api [None req-62918748-546f-4030-9046-6bb7ec988ab2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315532, 'name': PowerOffVM_Task, 'duration_secs': 0.253151} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.732785] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-62918748-546f-4030-9046-6bb7ec988ab2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 874.732980] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-62918748-546f-4030-9046-6bb7ec988ab2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 874.733260] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-80c96237-1ad5-42a7-b250-227112b76acd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.797204] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.798409] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-62918748-546f-4030-9046-6bb7ec988ab2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 874.798640] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-62918748-546f-4030-9046-6bb7ec988ab2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Deleting contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 874.798841] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-62918748-546f-4030-9046-6bb7ec988ab2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Deleting the datastore file [datastore2] 52a5afc6-4e79-436c-bc94-b61ca9fb860c {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 874.799464] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2c5b0943-3f53-4edc-bf8c-2c1772cece2e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.809539] env[61868]: DEBUG oslo_vmware.api [None req-62918748-546f-4030-9046-6bb7ec988ab2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Waiting for the task: (returnval){ [ 874.809539] env[61868]: value = "task-1315535" [ 874.809539] env[61868]: _type = "Task" [ 874.809539] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.811804] env[61868]: DEBUG oslo_vmware.api [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52464826-7b2d-57dd-4630-0d60519116f0, 'name': SearchDatastore_Task, 'duration_secs': 0.012615} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.815219] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.815545] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] c6108cf2-b597-4ca7-8f57-12ea82cc6c2c/c6108cf2-b597-4ca7-8f57-12ea82cc6c2c.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 874.815889] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.816159] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 874.816312] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8907e4ec-8ca5-4698-b035-b868cfebbf05 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.820039] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-77a9ec59-d0f8-4bac-b041-6fb2184738b8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.820913] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9d9434da-029e-4450-831a-50e8cce87561 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.827273] env[61868]: DEBUG oslo_vmware.api [None req-62918748-546f-4030-9046-6bb7ec988ab2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315535, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.829705] env[61868]: DEBUG oslo_vmware.api [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Waiting for the task: (returnval){ [ 874.829705] env[61868]: value = "task-1315536" [ 874.829705] env[61868]: _type = "Task" [ 874.829705] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.833182] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 874.833369] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 874.834604] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-88597fb0-df91-4f72-8865-f668db5834aa {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.841142] env[61868]: DEBUG oslo_vmware.api [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Task: {'id': task-1315536, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.844826] env[61868]: DEBUG oslo_vmware.api [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for the task: (returnval){ [ 874.844826] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52048ade-fe0f-3503-4912-a851f9b579d8" [ 874.844826] env[61868]: _type = "Task" [ 874.844826] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.853081] env[61868]: DEBUG oslo_vmware.api [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52048ade-fe0f-3503-4912-a851f9b579d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.011900] env[61868]: DEBUG oslo_vmware.api [None req-a76b1c97-d8a9-4316-9001-c2415aa6d377 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315533, 'name': PowerOffVM_Task, 'duration_secs': 0.308976} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.012481] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-a76b1c97-d8a9-4316-9001-c2415aa6d377 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 875.012669] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a76b1c97-d8a9-4316-9001-c2415aa6d377 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 875.012971] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4e72eaf6-4e92-4ac9-9f4d-859b9c9bfa19 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.117528] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a76b1c97-d8a9-4316-9001-c2415aa6d377 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 875.117877] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a76b1c97-d8a9-4316-9001-c2415aa6d377 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Deleting contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 875.117993] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-a76b1c97-d8a9-4316-9001-c2415aa6d377 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Deleting the datastore file [datastore2] c2693a43-4ea2-4ab0-8915-2fa544780e3c {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 875.118243] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-089833ce-2280-4987-816b-969be3656b18 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.125083] env[61868]: DEBUG oslo_concurrency.lockutils [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.494s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.125083] env[61868]: DEBUG nova.compute.manager [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 875.128650] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.695s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.128904] env[61868]: DEBUG nova.objects.instance [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lazy-loading 'resources' on Instance uuid a16a726b-9673-4de7-9188-19c608a7dc0b {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 875.131275] env[61868]: DEBUG oslo_vmware.api [None req-a76b1c97-d8a9-4316-9001-c2415aa6d377 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Waiting for the task: (returnval){ [ 875.131275] env[61868]: value = "task-1315538" [ 875.131275] env[61868]: _type = "Task" [ 875.131275] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.143664] env[61868]: DEBUG oslo_vmware.api [None req-a76b1c97-d8a9-4316-9001-c2415aa6d377 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315538, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.323577] env[61868]: DEBUG oslo_vmware.api [None req-62918748-546f-4030-9046-6bb7ec988ab2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315535, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162374} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.324109] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-62918748-546f-4030-9046-6bb7ec988ab2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 875.324366] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-62918748-546f-4030-9046-6bb7ec988ab2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Deleted contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 875.324659] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-62918748-546f-4030-9046-6bb7ec988ab2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 875.324949] env[61868]: INFO nova.compute.manager [None req-62918748-546f-4030-9046-6bb7ec988ab2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Took 1.12 seconds to destroy the instance on the hypervisor. [ 875.325341] env[61868]: DEBUG oslo.service.loopingcall [None req-62918748-546f-4030-9046-6bb7ec988ab2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 875.325650] env[61868]: DEBUG nova.compute.manager [-] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 875.325803] env[61868]: DEBUG nova.network.neutron [-] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 875.341442] env[61868]: DEBUG oslo_vmware.api [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Task: {'id': task-1315536, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.354855] env[61868]: DEBUG oslo_vmware.api [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52048ade-fe0f-3503-4912-a851f9b579d8, 'name': SearchDatastore_Task, 'duration_secs': 0.013266} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.355737] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c62bdd03-1da3-402c-8219-48001853e306 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.366943] env[61868]: DEBUG oslo_vmware.api [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for the task: (returnval){ [ 875.366943] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52cb761d-bff6-b881-898d-69043753fd29" [ 875.366943] env[61868]: _type = "Task" [ 875.366943] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.375089] env[61868]: DEBUG oslo_vmware.api [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52cb761d-bff6-b881-898d-69043753fd29, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.545520] env[61868]: DEBUG nova.compute.manager [req-fb368d87-bb96-4c3e-b134-a618f575bfc7 req-48aafd6f-887b-4626-922e-38c827a61a22 service nova] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Received event network-vif-plugged-ae9b8f9e-17ec-4c1f-8052-8467a711410c {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 875.545948] env[61868]: DEBUG oslo_concurrency.lockutils [req-fb368d87-bb96-4c3e-b134-a618f575bfc7 req-48aafd6f-887b-4626-922e-38c827a61a22 service nova] Acquiring lock "b9b5be37-6b30-4229-9c8d-3ee9d30db119-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.546301] env[61868]: DEBUG oslo_concurrency.lockutils [req-fb368d87-bb96-4c3e-b134-a618f575bfc7 req-48aafd6f-887b-4626-922e-38c827a61a22 service nova] Lock "b9b5be37-6b30-4229-9c8d-3ee9d30db119-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.546587] env[61868]: DEBUG oslo_concurrency.lockutils [req-fb368d87-bb96-4c3e-b134-a618f575bfc7 req-48aafd6f-887b-4626-922e-38c827a61a22 service nova] Lock "b9b5be37-6b30-4229-9c8d-3ee9d30db119-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.546873] env[61868]: DEBUG nova.compute.manager [req-fb368d87-bb96-4c3e-b134-a618f575bfc7 req-48aafd6f-887b-4626-922e-38c827a61a22 service nova] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] No waiting events found dispatching network-vif-plugged-ae9b8f9e-17ec-4c1f-8052-8467a711410c {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 875.547173] env[61868]: WARNING nova.compute.manager [req-fb368d87-bb96-4c3e-b134-a618f575bfc7 req-48aafd6f-887b-4626-922e-38c827a61a22 service nova] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Received unexpected event network-vif-plugged-ae9b8f9e-17ec-4c1f-8052-8467a711410c for instance with vm_state building and task_state spawning. [ 875.580047] env[61868]: DEBUG nova.network.neutron [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Successfully updated port: ae9b8f9e-17ec-4c1f-8052-8467a711410c {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 875.631744] env[61868]: DEBUG nova.objects.instance [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lazy-loading 'numa_topology' on Instance uuid a16a726b-9673-4de7-9188-19c608a7dc0b {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 875.633837] env[61868]: DEBUG nova.compute.utils [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 875.635613] env[61868]: DEBUG nova.compute.manager [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 875.636149] env[61868]: DEBUG nova.network.neutron [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 875.655178] env[61868]: DEBUG oslo_vmware.api [None req-a76b1c97-d8a9-4316-9001-c2415aa6d377 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Task: {'id': task-1315538, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.415902} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.656597] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-a76b1c97-d8a9-4316-9001-c2415aa6d377 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 875.656597] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a76b1c97-d8a9-4316-9001-c2415aa6d377 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Deleted contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 875.656597] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a76b1c97-d8a9-4316-9001-c2415aa6d377 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 875.658389] env[61868]: INFO nova.compute.manager [None req-a76b1c97-d8a9-4316-9001-c2415aa6d377 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Took 1.18 seconds to destroy the instance on the hypervisor. [ 875.658472] env[61868]: DEBUG oslo.service.loopingcall [None req-a76b1c97-d8a9-4316-9001-c2415aa6d377 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 875.659444] env[61868]: DEBUG nova.compute.manager [-] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 875.659644] env[61868]: DEBUG nova.network.neutron [-] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 875.718682] env[61868]: DEBUG nova.policy [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'daad105e7edf4fb0ae0b2e685bfd92e9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b520c3ac58074e8d9b0bfafb817244a8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 875.840332] env[61868]: DEBUG oslo_vmware.api [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Task: {'id': task-1315536, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.513588} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.840662] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] c6108cf2-b597-4ca7-8f57-12ea82cc6c2c/c6108cf2-b597-4ca7-8f57-12ea82cc6c2c.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 875.840884] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 875.841216] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-23651a36-cbb7-4211-9b7a-badbdc24db90 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.847586] env[61868]: DEBUG oslo_vmware.api [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Waiting for the task: (returnval){ [ 875.847586] env[61868]: value = "task-1315539" [ 875.847586] env[61868]: _type = "Task" [ 875.847586] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.855723] env[61868]: DEBUG oslo_vmware.api [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Task: {'id': task-1315539, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.876646] env[61868]: DEBUG oslo_vmware.api [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52cb761d-bff6-b881-898d-69043753fd29, 'name': SearchDatastore_Task, 'duration_secs': 0.012518} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.876819] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.877045] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 5aec2609-79d3-4725-a182-70b575adbe17/5aec2609-79d3-4725-a182-70b575adbe17.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 875.877403] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-72280640-5bc4-4f7a-8c1e-f8b5463fa62b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.883461] env[61868]: DEBUG oslo_vmware.api [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for the task: (returnval){ [ 875.883461] env[61868]: value = "task-1315540" [ 875.883461] env[61868]: _type = "Task" [ 875.883461] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.891409] env[61868]: DEBUG oslo_vmware.api [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315540, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.081776] env[61868]: DEBUG oslo_concurrency.lockutils [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Acquiring lock "refresh_cache-b9b5be37-6b30-4229-9c8d-3ee9d30db119" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.081776] env[61868]: DEBUG oslo_concurrency.lockutils [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Acquired lock "refresh_cache-b9b5be37-6b30-4229-9c8d-3ee9d30db119" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.081776] env[61868]: DEBUG nova.network.neutron [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 876.134853] env[61868]: DEBUG nova.network.neutron [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Successfully created port: 8ca7ca99-eddc-41fa-bfbc-1fafcaf733cc {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 876.137092] env[61868]: DEBUG nova.objects.base [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=61868) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 876.146137] env[61868]: DEBUG nova.compute.manager [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 876.223467] env[61868]: DEBUG nova.network.neutron [-] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.359970] env[61868]: DEBUG oslo_vmware.api [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Task: {'id': task-1315539, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074073} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.363181] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 876.364764] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02cc8a05-302e-4e27-acbf-b14ffca50198 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.388519] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] c6108cf2-b597-4ca7-8f57-12ea82cc6c2c/c6108cf2-b597-4ca7-8f57-12ea82cc6c2c.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 876.392252] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2e9826ab-f35f-4ce3-b172-d35f0c2fe5de {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.416166] env[61868]: DEBUG oslo_vmware.api [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315540, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.497273} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.417145] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 5aec2609-79d3-4725-a182-70b575adbe17/5aec2609-79d3-4725-a182-70b575adbe17.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 876.418018] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 876.418018] env[61868]: DEBUG oslo_vmware.api [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Waiting for the task: (returnval){ [ 876.418018] env[61868]: value = "task-1315541" [ 876.418018] env[61868]: _type = "Task" [ 876.418018] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.420327] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c457098c-bc2d-4d4a-9806-cc23789a5e08 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.431705] env[61868]: DEBUG oslo_vmware.api [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for the task: (returnval){ [ 876.431705] env[61868]: value = "task-1315542" [ 876.431705] env[61868]: _type = "Task" [ 876.431705] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.432622] env[61868]: DEBUG oslo_vmware.api [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Task: {'id': task-1315541, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.445705] env[61868]: DEBUG oslo_vmware.api [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315542, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.520923] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43fc6d7f-60a7-4414-a9be-536d5593c02a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.529078] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ac00816-c146-4722-a8ac-3a4da2173597 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.564036] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-548e04c6-f02d-4bfa-a691-244cdd851989 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.572786] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2924bc44-1293-42ed-9b47-8f0c13bc5b83 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.588563] env[61868]: DEBUG nova.network.neutron [-] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.591969] env[61868]: DEBUG nova.compute.provider_tree [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 876.609096] env[61868]: DEBUG nova.compute.manager [req-bfa320e9-bbb6-4259-9ef8-6212aab6d640 req-4ce1847d-1497-44cb-9edf-28f3aea00fa6 service nova] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Received event network-vif-deleted-de6238cd-1354-4052-9aff-0d28691de0c8 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 876.616084] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Acquiring lock "b51e968e-6911-47f3-b5a9-e8a5865e42a0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.616084] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Lock "b51e968e-6911-47f3-b5a9-e8a5865e42a0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.629981] env[61868]: DEBUG nova.network.neutron [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 876.727678] env[61868]: INFO nova.compute.manager [-] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Took 1.40 seconds to deallocate network for instance. [ 876.768887] env[61868]: DEBUG nova.network.neutron [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Updating instance_info_cache with network_info: [{"id": "ae9b8f9e-17ec-4c1f-8052-8467a711410c", "address": "fa:16:3e:d5:7e:6e", "network": {"id": "a8d3ac26-3753-41a8-bddb-147d4cdb4934", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1451108424-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e313c27c7d7342e3bc6501653c0c685d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae9b8f9e-17", "ovs_interfaceid": "ae9b8f9e-17ec-4c1f-8052-8467a711410c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.931515] env[61868]: DEBUG oslo_vmware.api [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Task: {'id': task-1315541, 'name': ReconfigVM_Task, 'duration_secs': 0.26834} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.931833] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Reconfigured VM instance instance-00000047 to attach disk [datastore2] c6108cf2-b597-4ca7-8f57-12ea82cc6c2c/c6108cf2-b597-4ca7-8f57-12ea82cc6c2c.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 876.932550] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-929dbc8a-dca4-4558-8126-67e3272412e0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.941518] env[61868]: DEBUG oslo_vmware.api [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315542, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06908} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.942712] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 876.943088] env[61868]: DEBUG oslo_vmware.api [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Waiting for the task: (returnval){ [ 876.943088] env[61868]: value = "task-1315543" [ 876.943088] env[61868]: _type = "Task" [ 876.943088] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.943746] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac8713ae-b0af-461c-a624-cb18cb2f69ed {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.968600] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] 5aec2609-79d3-4725-a182-70b575adbe17/5aec2609-79d3-4725-a182-70b575adbe17.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 876.971897] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-83926196-6298-4bca-9a16-57f3b8cba1ae {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.986027] env[61868]: DEBUG oslo_vmware.api [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Task: {'id': task-1315543, 'name': Rename_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.990753] env[61868]: DEBUG oslo_vmware.api [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for the task: (returnval){ [ 876.990753] env[61868]: value = "task-1315544" [ 876.990753] env[61868]: _type = "Task" [ 876.990753] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.998375] env[61868]: DEBUG oslo_vmware.api [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315544, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.094016] env[61868]: INFO nova.compute.manager [-] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Took 1.43 seconds to deallocate network for instance. [ 877.096434] env[61868]: DEBUG nova.scheduler.client.report [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 877.156050] env[61868]: DEBUG nova.compute.manager [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 877.182977] env[61868]: DEBUG nova.virt.hardware [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 877.183277] env[61868]: DEBUG nova.virt.hardware [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 877.183442] env[61868]: DEBUG nova.virt.hardware [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 877.183629] env[61868]: DEBUG nova.virt.hardware [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 877.183782] env[61868]: DEBUG nova.virt.hardware [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 877.183935] env[61868]: DEBUG nova.virt.hardware [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 877.184208] env[61868]: DEBUG nova.virt.hardware [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 877.184382] env[61868]: DEBUG nova.virt.hardware [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 877.184554] env[61868]: DEBUG nova.virt.hardware [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 877.184741] env[61868]: DEBUG nova.virt.hardware [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 877.184923] env[61868]: DEBUG nova.virt.hardware [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 877.185811] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12313c64-4ad4-4871-898a-8e13d9f655c1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.194677] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61ba95bb-cbc9-4345-9b1b-a4f0a3529e8c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.233971] env[61868]: DEBUG oslo_concurrency.lockutils [None req-62918748-546f-4030-9046-6bb7ec988ab2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.272313] env[61868]: DEBUG oslo_concurrency.lockutils [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Releasing lock "refresh_cache-b9b5be37-6b30-4229-9c8d-3ee9d30db119" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.272657] env[61868]: DEBUG nova.compute.manager [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Instance network_info: |[{"id": "ae9b8f9e-17ec-4c1f-8052-8467a711410c", "address": "fa:16:3e:d5:7e:6e", "network": {"id": "a8d3ac26-3753-41a8-bddb-147d4cdb4934", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1451108424-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e313c27c7d7342e3bc6501653c0c685d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae9b8f9e-17", "ovs_interfaceid": "ae9b8f9e-17ec-4c1f-8052-8467a711410c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 877.273209] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d5:7e:6e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dc6d5964-1106-4345-a26d-185dabd4ff0f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ae9b8f9e-17ec-4c1f-8052-8467a711410c', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 877.281841] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Creating folder: Project (e313c27c7d7342e3bc6501653c0c685d). Parent ref: group-v281478. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 877.282168] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1d49c3b9-8d71-46ef-a6e6-e06ff52df4dd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.292315] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Created folder: Project (e313c27c7d7342e3bc6501653c0c685d) in parent group-v281478. [ 877.292558] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Creating folder: Instances. Parent ref: group-v281548. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 877.292832] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-da79e56e-bea4-4abc-ab9e-0c08423da9a6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.301370] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Created folder: Instances in parent group-v281548. [ 877.301616] env[61868]: DEBUG oslo.service.loopingcall [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 877.301904] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 877.302115] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-15a415b7-6de0-4b9b-aaac-d6eeceddc394 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.323026] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 877.323026] env[61868]: value = "task-1315547" [ 877.323026] env[61868]: _type = "Task" [ 877.323026] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.329434] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315547, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.455139] env[61868]: DEBUG oslo_vmware.api [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Task: {'id': task-1315543, 'name': Rename_Task, 'duration_secs': 0.132476} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.455414] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 877.455661] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fea26bfd-d6a7-4158-bcb0-de765377587d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.462523] env[61868]: DEBUG oslo_vmware.api [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Waiting for the task: (returnval){ [ 877.462523] env[61868]: value = "task-1315548" [ 877.462523] env[61868]: _type = "Task" [ 877.462523] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.470608] env[61868]: DEBUG oslo_vmware.api [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Task: {'id': task-1315548, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.501167] env[61868]: DEBUG oslo_vmware.api [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315544, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.605429] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.475s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.610165] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a76b1c97-d8a9-4316-9001-c2415aa6d377 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.610826] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.064s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.612770] env[61868]: INFO nova.compute.claims [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 877.654917] env[61868]: DEBUG nova.compute.manager [req-1c8ec920-643f-47b0-838f-f2f7ec9447af req-f687408f-d2aa-4967-931e-6085f3ca9fd6 service nova] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Received event network-changed-ae9b8f9e-17ec-4c1f-8052-8467a711410c {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 877.655152] env[61868]: DEBUG nova.compute.manager [req-1c8ec920-643f-47b0-838f-f2f7ec9447af req-f687408f-d2aa-4967-931e-6085f3ca9fd6 service nova] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Refreshing instance network info cache due to event network-changed-ae9b8f9e-17ec-4c1f-8052-8467a711410c. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 877.655374] env[61868]: DEBUG oslo_concurrency.lockutils [req-1c8ec920-643f-47b0-838f-f2f7ec9447af req-f687408f-d2aa-4967-931e-6085f3ca9fd6 service nova] Acquiring lock "refresh_cache-b9b5be37-6b30-4229-9c8d-3ee9d30db119" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.655483] env[61868]: DEBUG oslo_concurrency.lockutils [req-1c8ec920-643f-47b0-838f-f2f7ec9447af req-f687408f-d2aa-4967-931e-6085f3ca9fd6 service nova] Acquired lock "refresh_cache-b9b5be37-6b30-4229-9c8d-3ee9d30db119" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.655653] env[61868]: DEBUG nova.network.neutron [req-1c8ec920-643f-47b0-838f-f2f7ec9447af req-f687408f-d2aa-4967-931e-6085f3ca9fd6 service nova] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Refreshing network info cache for port ae9b8f9e-17ec-4c1f-8052-8467a711410c {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 877.717203] env[61868]: DEBUG nova.network.neutron [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Successfully updated port: 8ca7ca99-eddc-41fa-bfbc-1fafcaf733cc {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 877.832575] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315547, 'name': CreateVM_Task, 'duration_secs': 0.398806} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.832877] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 877.833682] env[61868]: DEBUG oslo_concurrency.lockutils [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.833910] env[61868]: DEBUG oslo_concurrency.lockutils [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.834307] env[61868]: DEBUG oslo_concurrency.lockutils [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 877.834714] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ae77219-2847-41e2-9686-6beb188f7666 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.839353] env[61868]: DEBUG oslo_vmware.api [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Waiting for the task: (returnval){ [ 877.839353] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]520fb973-65a3-26cb-ffa5-969422157c3a" [ 877.839353] env[61868]: _type = "Task" [ 877.839353] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.849171] env[61868]: DEBUG oslo_vmware.api [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]520fb973-65a3-26cb-ffa5-969422157c3a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.975986] env[61868]: DEBUG oslo_vmware.api [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Task: {'id': task-1315548, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.003346] env[61868]: DEBUG oslo_vmware.api [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315544, 'name': ReconfigVM_Task, 'duration_secs': 0.732524} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.003647] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Reconfigured VM instance instance-00000046 to attach disk [datastore2] 5aec2609-79d3-4725-a182-70b575adbe17/5aec2609-79d3-4725-a182-70b575adbe17.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 878.004318] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c6357155-073e-48e9-a87e-cddde45deb56 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.010129] env[61868]: DEBUG oslo_vmware.api [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for the task: (returnval){ [ 878.010129] env[61868]: value = "task-1315549" [ 878.010129] env[61868]: _type = "Task" [ 878.010129] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.017814] env[61868]: DEBUG oslo_vmware.api [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315549, 'name': Rename_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.119225] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c7568b0c-dcaf-4be9-88e4-6407e0537d51 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "a16a726b-9673-4de7-9188-19c608a7dc0b" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 39.540s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.120322] env[61868]: DEBUG oslo_concurrency.lockutils [None req-75a27b94-21b6-4932-bd97-f47ebe7e57ec tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "a16a726b-9673-4de7-9188-19c608a7dc0b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 17.992s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.120604] env[61868]: DEBUG oslo_concurrency.lockutils [None req-75a27b94-21b6-4932-bd97-f47ebe7e57ec tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "a16a726b-9673-4de7-9188-19c608a7dc0b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.120958] env[61868]: DEBUG oslo_concurrency.lockutils [None req-75a27b94-21b6-4932-bd97-f47ebe7e57ec tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "a16a726b-9673-4de7-9188-19c608a7dc0b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.121158] env[61868]: DEBUG oslo_concurrency.lockutils [None req-75a27b94-21b6-4932-bd97-f47ebe7e57ec tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "a16a726b-9673-4de7-9188-19c608a7dc0b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.124289] env[61868]: INFO nova.compute.manager [None req-75a27b94-21b6-4932-bd97-f47ebe7e57ec tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Terminating instance [ 878.126565] env[61868]: DEBUG nova.compute.manager [None req-75a27b94-21b6-4932-bd97-f47ebe7e57ec tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 878.126763] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-75a27b94-21b6-4932-bd97-f47ebe7e57ec tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 878.127037] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d0a88096-3c25-469d-8b30-653b9a515043 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.137090] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ed61a50-22f7-473f-8a52-9e297cf0dbfd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.165283] env[61868]: WARNING nova.virt.vmwareapi.vmops [None req-75a27b94-21b6-4932-bd97-f47ebe7e57ec tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a16a726b-9673-4de7-9188-19c608a7dc0b could not be found. [ 878.165516] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-75a27b94-21b6-4932-bd97-f47ebe7e57ec tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 878.165690] env[61868]: INFO nova.compute.manager [None req-75a27b94-21b6-4932-bd97-f47ebe7e57ec tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 878.165931] env[61868]: DEBUG oslo.service.loopingcall [None req-75a27b94-21b6-4932-bd97-f47ebe7e57ec tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 878.166193] env[61868]: DEBUG nova.compute.manager [-] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 878.166288] env[61868]: DEBUG nova.network.neutron [-] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 878.220105] env[61868]: DEBUG oslo_concurrency.lockutils [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "refresh_cache-b7055e5a-37d2-42d5-bab0-1e70faf52b58" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.220262] env[61868]: DEBUG oslo_concurrency.lockutils [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquired lock "refresh_cache-b7055e5a-37d2-42d5-bab0-1e70faf52b58" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.220414] env[61868]: DEBUG nova.network.neutron [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 878.349969] env[61868]: DEBUG oslo_vmware.api [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]520fb973-65a3-26cb-ffa5-969422157c3a, 'name': SearchDatastore_Task, 'duration_secs': 0.0099} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.350298] env[61868]: DEBUG oslo_concurrency.lockutils [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.350567] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 878.350805] env[61868]: DEBUG oslo_concurrency.lockutils [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.350956] env[61868]: DEBUG oslo_concurrency.lockutils [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.351151] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 878.351407] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2fe70943-dd4b-4590-a6ee-4078e1bcc902 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.364441] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 878.364624] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 878.365355] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c5caac66-4e09-4227-b3a0-f4c8777f4ff7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.371218] env[61868]: DEBUG oslo_vmware.api [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Waiting for the task: (returnval){ [ 878.371218] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]529bf83e-08bd-4772-fd40-4fea5c5d8928" [ 878.371218] env[61868]: _type = "Task" [ 878.371218] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.381213] env[61868]: DEBUG oslo_vmware.api [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]529bf83e-08bd-4772-fd40-4fea5c5d8928, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.472366] env[61868]: DEBUG oslo_vmware.api [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Task: {'id': task-1315548, 'name': PowerOnVM_Task, 'duration_secs': 0.513929} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.472734] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 878.472994] env[61868]: INFO nova.compute.manager [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Took 8.32 seconds to spawn the instance on the hypervisor. [ 878.473438] env[61868]: DEBUG nova.compute.manager [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 878.474233] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bc530fe-f54a-41ef-a62f-b5af0a9a5b3f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.485869] env[61868]: DEBUG nova.network.neutron [req-1c8ec920-643f-47b0-838f-f2f7ec9447af req-f687408f-d2aa-4967-931e-6085f3ca9fd6 service nova] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Updated VIF entry in instance network info cache for port ae9b8f9e-17ec-4c1f-8052-8467a711410c. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 878.486261] env[61868]: DEBUG nova.network.neutron [req-1c8ec920-643f-47b0-838f-f2f7ec9447af req-f687408f-d2aa-4967-931e-6085f3ca9fd6 service nova] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Updating instance_info_cache with network_info: [{"id": "ae9b8f9e-17ec-4c1f-8052-8467a711410c", "address": "fa:16:3e:d5:7e:6e", "network": {"id": "a8d3ac26-3753-41a8-bddb-147d4cdb4934", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1451108424-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e313c27c7d7342e3bc6501653c0c685d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae9b8f9e-17", "ovs_interfaceid": "ae9b8f9e-17ec-4c1f-8052-8467a711410c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.520633] env[61868]: DEBUG oslo_vmware.api [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315549, 'name': Rename_Task, 'duration_secs': 0.186151} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.521602] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 878.522081] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3dcb6411-b8a1-4884-b72b-0567616a45e6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.529688] env[61868]: DEBUG oslo_vmware.api [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for the task: (returnval){ [ 878.529688] env[61868]: value = "task-1315550" [ 878.529688] env[61868]: _type = "Task" [ 878.529688] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.539295] env[61868]: DEBUG oslo_vmware.api [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315550, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.763623] env[61868]: DEBUG nova.network.neutron [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 878.882934] env[61868]: DEBUG oslo_vmware.api [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]529bf83e-08bd-4772-fd40-4fea5c5d8928, 'name': SearchDatastore_Task, 'duration_secs': 0.017683} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.883766] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-59ec1950-3853-4e94-ac66-a7206e663c37 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.891566] env[61868]: DEBUG oslo_vmware.api [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Waiting for the task: (returnval){ [ 878.891566] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52783bef-b37b-fd08-bb28-128f0c7f2cd8" [ 878.891566] env[61868]: _type = "Task" [ 878.891566] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.899238] env[61868]: DEBUG oslo_vmware.api [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52783bef-b37b-fd08-bb28-128f0c7f2cd8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.916198] env[61868]: DEBUG nova.network.neutron [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Updating instance_info_cache with network_info: [{"id": "8ca7ca99-eddc-41fa-bfbc-1fafcaf733cc", "address": "fa:16:3e:1b:0e:cb", "network": {"id": "77f33b52-e4eb-4b72-85d4-5991a2b5c36d", "bridge": "br-int", "label": "tempest-ServersTestJSON-748635995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b520c3ac58074e8d9b0bfafb817244a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ca7ca99-ed", "ovs_interfaceid": "8ca7ca99-eddc-41fa-bfbc-1fafcaf733cc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.928225] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b59a151-1b22-44a4-bb05-d2bdf81267d7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.936130] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e558c668-ca89-4511-b30d-741115f1af1b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.967857] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73856aeb-0b70-4759-b99d-cb6a618abe4a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.974344] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3959ce9c-8fa1-4feb-989a-f5a7eece72a1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.991040] env[61868]: DEBUG nova.compute.provider_tree [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 878.993741] env[61868]: DEBUG oslo_concurrency.lockutils [req-1c8ec920-643f-47b0-838f-f2f7ec9447af req-f687408f-d2aa-4967-931e-6085f3ca9fd6 service nova] Releasing lock "refresh_cache-b9b5be37-6b30-4229-9c8d-3ee9d30db119" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.993973] env[61868]: DEBUG nova.compute.manager [req-1c8ec920-643f-47b0-838f-f2f7ec9447af req-f687408f-d2aa-4967-931e-6085f3ca9fd6 service nova] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Received event network-vif-deleted-849a9665-e0b7-4a35-a860-0ff32923cb6c {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 878.994999] env[61868]: INFO nova.compute.manager [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Took 27.91 seconds to build instance. [ 879.035180] env[61868]: DEBUG nova.network.neutron [-] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.042909] env[61868]: DEBUG oslo_vmware.api [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315550, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.403707] env[61868]: DEBUG oslo_vmware.api [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52783bef-b37b-fd08-bb28-128f0c7f2cd8, 'name': SearchDatastore_Task, 'duration_secs': 0.03259} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.404024] env[61868]: DEBUG oslo_concurrency.lockutils [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.404287] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] b9b5be37-6b30-4229-9c8d-3ee9d30db119/b9b5be37-6b30-4229-9c8d-3ee9d30db119.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 879.404558] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7433f371-7eab-4b8c-b638-3df85420a64c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.411374] env[61868]: DEBUG oslo_vmware.api [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Waiting for the task: (returnval){ [ 879.411374] env[61868]: value = "task-1315551" [ 879.411374] env[61868]: _type = "Task" [ 879.411374] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.421225] env[61868]: DEBUG oslo_concurrency.lockutils [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Releasing lock "refresh_cache-b7055e5a-37d2-42d5-bab0-1e70faf52b58" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.421540] env[61868]: DEBUG nova.compute.manager [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Instance network_info: |[{"id": "8ca7ca99-eddc-41fa-bfbc-1fafcaf733cc", "address": "fa:16:3e:1b:0e:cb", "network": {"id": "77f33b52-e4eb-4b72-85d4-5991a2b5c36d", "bridge": "br-int", "label": "tempest-ServersTestJSON-748635995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b520c3ac58074e8d9b0bfafb817244a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ca7ca99-ed", "ovs_interfaceid": "8ca7ca99-eddc-41fa-bfbc-1fafcaf733cc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 879.421850] env[61868]: DEBUG oslo_vmware.api [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Task: {'id': task-1315551, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.422475] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1b:0e:cb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56136ef6-99d7-4562-9a9f-d66fec951c5c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8ca7ca99-eddc-41fa-bfbc-1fafcaf733cc', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 879.429800] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Creating folder: Project (b520c3ac58074e8d9b0bfafb817244a8). Parent ref: group-v281478. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 879.430093] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2b30ad36-807a-40ba-8cce-e9ce77434d62 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.440495] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Created folder: Project (b520c3ac58074e8d9b0bfafb817244a8) in parent group-v281478. [ 879.440680] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Creating folder: Instances. Parent ref: group-v281551. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 879.440905] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fb2de97e-f1e4-4b41-a10b-ab5380c1293f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.449284] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Created folder: Instances in parent group-v281551. [ 879.449606] env[61868]: DEBUG oslo.service.loopingcall [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 879.449863] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 879.450148] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-82b103ab-a7db-48a5-9a86-b383e0dc647a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.471427] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 879.471427] env[61868]: value = "task-1315554" [ 879.471427] env[61868]: _type = "Task" [ 879.471427] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.480287] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315554, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.496446] env[61868]: DEBUG nova.scheduler.client.report [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 879.499928] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1ef58f67-05d1-43cd-8930-a7290d02235f tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Lock "c6108cf2-b597-4ca7-8f57-12ea82cc6c2c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 82.607s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.541822] env[61868]: INFO nova.compute.manager [-] [instance: a16a726b-9673-4de7-9188-19c608a7dc0b] Took 1.37 seconds to deallocate network for instance. [ 879.541995] env[61868]: DEBUG oslo_vmware.api [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315550, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.800388] env[61868]: DEBUG nova.compute.manager [req-e39b695c-bcac-47cc-bdbe-1b98cac6f712 req-de3cea11-72d5-4347-a11c-ebcfb24cd72c service nova] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Received event network-vif-plugged-8ca7ca99-eddc-41fa-bfbc-1fafcaf733cc {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 879.800388] env[61868]: DEBUG oslo_concurrency.lockutils [req-e39b695c-bcac-47cc-bdbe-1b98cac6f712 req-de3cea11-72d5-4347-a11c-ebcfb24cd72c service nova] Acquiring lock "b7055e5a-37d2-42d5-bab0-1e70faf52b58-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.800653] env[61868]: DEBUG oslo_concurrency.lockutils [req-e39b695c-bcac-47cc-bdbe-1b98cac6f712 req-de3cea11-72d5-4347-a11c-ebcfb24cd72c service nova] Lock "b7055e5a-37d2-42d5-bab0-1e70faf52b58-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.801013] env[61868]: DEBUG oslo_concurrency.lockutils [req-e39b695c-bcac-47cc-bdbe-1b98cac6f712 req-de3cea11-72d5-4347-a11c-ebcfb24cd72c service nova] Lock "b7055e5a-37d2-42d5-bab0-1e70faf52b58-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.801448] env[61868]: DEBUG nova.compute.manager [req-e39b695c-bcac-47cc-bdbe-1b98cac6f712 req-de3cea11-72d5-4347-a11c-ebcfb24cd72c service nova] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] No waiting events found dispatching network-vif-plugged-8ca7ca99-eddc-41fa-bfbc-1fafcaf733cc {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 879.801800] env[61868]: WARNING nova.compute.manager [req-e39b695c-bcac-47cc-bdbe-1b98cac6f712 req-de3cea11-72d5-4347-a11c-ebcfb24cd72c service nova] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Received unexpected event network-vif-plugged-8ca7ca99-eddc-41fa-bfbc-1fafcaf733cc for instance with vm_state building and task_state spawning. [ 879.802116] env[61868]: DEBUG nova.compute.manager [req-e39b695c-bcac-47cc-bdbe-1b98cac6f712 req-de3cea11-72d5-4347-a11c-ebcfb24cd72c service nova] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Received event network-changed-8ca7ca99-eddc-41fa-bfbc-1fafcaf733cc {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 879.802399] env[61868]: DEBUG nova.compute.manager [req-e39b695c-bcac-47cc-bdbe-1b98cac6f712 req-de3cea11-72d5-4347-a11c-ebcfb24cd72c service nova] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Refreshing instance network info cache due to event network-changed-8ca7ca99-eddc-41fa-bfbc-1fafcaf733cc. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 879.802710] env[61868]: DEBUG oslo_concurrency.lockutils [req-e39b695c-bcac-47cc-bdbe-1b98cac6f712 req-de3cea11-72d5-4347-a11c-ebcfb24cd72c service nova] Acquiring lock "refresh_cache-b7055e5a-37d2-42d5-bab0-1e70faf52b58" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.803012] env[61868]: DEBUG oslo_concurrency.lockutils [req-e39b695c-bcac-47cc-bdbe-1b98cac6f712 req-de3cea11-72d5-4347-a11c-ebcfb24cd72c service nova] Acquired lock "refresh_cache-b7055e5a-37d2-42d5-bab0-1e70faf52b58" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.803311] env[61868]: DEBUG nova.network.neutron [req-e39b695c-bcac-47cc-bdbe-1b98cac6f712 req-de3cea11-72d5-4347-a11c-ebcfb24cd72c service nova] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Refreshing network info cache for port 8ca7ca99-eddc-41fa-bfbc-1fafcaf733cc {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 879.920523] env[61868]: DEBUG oslo_vmware.api [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Task: {'id': task-1315551, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.469109} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.921030] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] b9b5be37-6b30-4229-9c8d-3ee9d30db119/b9b5be37-6b30-4229-9c8d-3ee9d30db119.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 879.921415] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 879.923716] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c4b7afd6-16f1-4673-aa3f-58ca18788329 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.931025] env[61868]: DEBUG oslo_vmware.api [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Waiting for the task: (returnval){ [ 879.931025] env[61868]: value = "task-1315555" [ 879.931025] env[61868]: _type = "Task" [ 879.931025] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.938599] env[61868]: DEBUG oslo_vmware.api [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Task: {'id': task-1315555, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.981766] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315554, 'name': CreateVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.004478] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.392s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.004478] env[61868]: DEBUG nova.compute.manager [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 880.006532] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0a2a4111-8de4-4541-8ce5-040de2bcae6b tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.367s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.006532] env[61868]: DEBUG nova.objects.instance [None req-0a2a4111-8de4-4541-8ce5-040de2bcae6b tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Lazy-loading 'resources' on Instance uuid ca85dfd1-e794-41dc-a337-fbd99998e1fa {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 880.007614] env[61868]: DEBUG nova.compute.manager [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 880.043159] env[61868]: DEBUG oslo_vmware.api [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315550, 'name': PowerOnVM_Task, 'duration_secs': 1.044178} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.043521] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 880.043627] env[61868]: INFO nova.compute.manager [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Took 12.41 seconds to spawn the instance on the hypervisor. [ 880.043805] env[61868]: DEBUG nova.compute.manager [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 880.044867] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e935c257-f3cf-468c-b88f-88d4c8fc2395 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.325537] env[61868]: DEBUG nova.compute.manager [req-35c48f24-7bbf-496d-9486-2481f2133676 req-05989015-6cfa-44a0-9865-8829f2f05c62 service nova] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Received event network-changed-981ca2e3-c2d7-4d28-bda6-68ddd6d57bc5 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 880.325785] env[61868]: DEBUG nova.compute.manager [req-35c48f24-7bbf-496d-9486-2481f2133676 req-05989015-6cfa-44a0-9865-8829f2f05c62 service nova] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Refreshing instance network info cache due to event network-changed-981ca2e3-c2d7-4d28-bda6-68ddd6d57bc5. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 880.326389] env[61868]: DEBUG oslo_concurrency.lockutils [req-35c48f24-7bbf-496d-9486-2481f2133676 req-05989015-6cfa-44a0-9865-8829f2f05c62 service nova] Acquiring lock "refresh_cache-c6108cf2-b597-4ca7-8f57-12ea82cc6c2c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.326604] env[61868]: DEBUG oslo_concurrency.lockutils [req-35c48f24-7bbf-496d-9486-2481f2133676 req-05989015-6cfa-44a0-9865-8829f2f05c62 service nova] Acquired lock "refresh_cache-c6108cf2-b597-4ca7-8f57-12ea82cc6c2c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.326865] env[61868]: DEBUG nova.network.neutron [req-35c48f24-7bbf-496d-9486-2481f2133676 req-05989015-6cfa-44a0-9865-8829f2f05c62 service nova] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Refreshing network info cache for port 981ca2e3-c2d7-4d28-bda6-68ddd6d57bc5 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 880.438209] env[61868]: DEBUG oslo_vmware.api [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Task: {'id': task-1315555, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066039} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.438543] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 880.439337] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cde20fde-afd1-4414-b72d-57d785041234 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.462176] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] b9b5be37-6b30-4229-9c8d-3ee9d30db119/b9b5be37-6b30-4229-9c8d-3ee9d30db119.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 880.465509] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c6ee9b9b-1aed-4402-b4f7-88fe237f0100 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.493086] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315554, 'name': CreateVM_Task, 'duration_secs': 0.596291} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.494938] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 880.495337] env[61868]: DEBUG oslo_vmware.api [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Waiting for the task: (returnval){ [ 880.495337] env[61868]: value = "task-1315556" [ 880.495337] env[61868]: _type = "Task" [ 880.495337] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.495998] env[61868]: DEBUG oslo_concurrency.lockutils [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.496196] env[61868]: DEBUG oslo_concurrency.lockutils [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.496541] env[61868]: DEBUG oslo_concurrency.lockutils [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 880.496863] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa3e2405-7c87-4345-9046-e852873a6b11 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.508275] env[61868]: DEBUG oslo_vmware.api [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Task: {'id': task-1315556, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.510830] env[61868]: DEBUG nova.compute.utils [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 880.512730] env[61868]: DEBUG oslo_vmware.api [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 880.512730] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]523017bf-f3f9-fe46-90b1-3be7c603e965" [ 880.512730] env[61868]: _type = "Task" [ 880.512730] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.521345] env[61868]: DEBUG nova.compute.manager [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 880.521345] env[61868]: DEBUG nova.network.neutron [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 880.535803] env[61868]: DEBUG oslo_vmware.api [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]523017bf-f3f9-fe46-90b1-3be7c603e965, 'name': SearchDatastore_Task, 'duration_secs': 0.014469} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.536097] env[61868]: DEBUG oslo_concurrency.lockutils [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.536295] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 880.536549] env[61868]: DEBUG oslo_concurrency.lockutils [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.536765] env[61868]: DEBUG oslo_concurrency.lockutils [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.536902] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 880.537239] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a3306325-d22b-4d1d-aebb-4829ed22295a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.543745] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.555881] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 880.556318] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 880.558305] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bffcafb5-8a59-42c0-936a-80c05af85ad3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.579673] env[61868]: INFO nova.compute.manager [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Took 29.67 seconds to build instance. [ 880.583017] env[61868]: DEBUG oslo_vmware.api [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 880.583017] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]521ca089-c0fa-398e-9e9f-5552754e8b1a" [ 880.583017] env[61868]: _type = "Task" [ 880.583017] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.583017] env[61868]: DEBUG oslo_concurrency.lockutils [None req-75a27b94-21b6-4932-bd97-f47ebe7e57ec tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "a16a726b-9673-4de7-9188-19c608a7dc0b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.462s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.600874] env[61868]: DEBUG oslo_vmware.api [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]521ca089-c0fa-398e-9e9f-5552754e8b1a, 'name': SearchDatastore_Task, 'duration_secs': 0.014118} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.607243] env[61868]: DEBUG nova.network.neutron [req-e39b695c-bcac-47cc-bdbe-1b98cac6f712 req-de3cea11-72d5-4347-a11c-ebcfb24cd72c service nova] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Updated VIF entry in instance network info cache for port 8ca7ca99-eddc-41fa-bfbc-1fafcaf733cc. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 880.607702] env[61868]: DEBUG nova.network.neutron [req-e39b695c-bcac-47cc-bdbe-1b98cac6f712 req-de3cea11-72d5-4347-a11c-ebcfb24cd72c service nova] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Updating instance_info_cache with network_info: [{"id": "8ca7ca99-eddc-41fa-bfbc-1fafcaf733cc", "address": "fa:16:3e:1b:0e:cb", "network": {"id": "77f33b52-e4eb-4b72-85d4-5991a2b5c36d", "bridge": "br-int", "label": "tempest-ServersTestJSON-748635995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b520c3ac58074e8d9b0bfafb817244a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ca7ca99-ed", "ovs_interfaceid": "8ca7ca99-eddc-41fa-bfbc-1fafcaf733cc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.609198] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db712172-2282-41b2-9f4b-721975029304 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.615839] env[61868]: DEBUG nova.policy [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f9fbbb30d75e4ca1ac756825ec6d389a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '78a206c3daa14cc58e7fe8546f565ae0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 880.623041] env[61868]: DEBUG oslo_vmware.api [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 880.623041] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]521ce4c2-057e-d00d-2f02-4f09bb0f5665" [ 880.623041] env[61868]: _type = "Task" [ 880.623041] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.633086] env[61868]: DEBUG oslo_vmware.api [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]521ce4c2-057e-d00d-2f02-4f09bb0f5665, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.875106] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72024cbb-d6b7-4377-94a8-623f6c69eace {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.887940] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83e82ade-e18d-4abd-ba29-9fa04176d4bc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.920538] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b1b01ef-0a9f-44da-9db7-8bf4de222a21 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.929194] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40dda692-46fc-43a4-bdc1-23b8032dbdba {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.943467] env[61868]: DEBUG nova.compute.provider_tree [None req-0a2a4111-8de4-4541-8ce5-040de2bcae6b tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 881.005401] env[61868]: DEBUG oslo_vmware.api [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Task: {'id': task-1315556, 'name': ReconfigVM_Task, 'duration_secs': 0.367072} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.006823] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Reconfigured VM instance instance-00000048 to attach disk [datastore1] b9b5be37-6b30-4229-9c8d-3ee9d30db119/b9b5be37-6b30-4229-9c8d-3ee9d30db119.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 881.006823] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-57864462-e773-4001-968c-2a09b5d8c53d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.012477] env[61868]: DEBUG oslo_vmware.api [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Waiting for the task: (returnval){ [ 881.012477] env[61868]: value = "task-1315557" [ 881.012477] env[61868]: _type = "Task" [ 881.012477] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.015645] env[61868]: DEBUG nova.compute.manager [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 881.023090] env[61868]: DEBUG oslo_vmware.api [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Task: {'id': task-1315557, 'name': Rename_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.058152] env[61868]: DEBUG oslo_concurrency.lockutils [None req-be680e47-2395-47e2-bfb6-da2956f42022 tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Acquiring lock "c6108cf2-b597-4ca7-8f57-12ea82cc6c2c" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.058624] env[61868]: DEBUG oslo_concurrency.lockutils [None req-be680e47-2395-47e2-bfb6-da2956f42022 tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Lock "c6108cf2-b597-4ca7-8f57-12ea82cc6c2c" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.058624] env[61868]: INFO nova.compute.manager [None req-be680e47-2395-47e2-bfb6-da2956f42022 tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Rebooting instance [ 881.091887] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0dad6915-d531-430f-b96c-e150e4c58bff tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "5aec2609-79d3-4725-a182-70b575adbe17" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 87.374s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.117087] env[61868]: DEBUG oslo_concurrency.lockutils [req-e39b695c-bcac-47cc-bdbe-1b98cac6f712 req-de3cea11-72d5-4347-a11c-ebcfb24cd72c service nova] Releasing lock "refresh_cache-b7055e5a-37d2-42d5-bab0-1e70faf52b58" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.119825] env[61868]: DEBUG nova.network.neutron [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Successfully created port: 29e8f7bd-db77-4071-b4c6-79e8338c2e82 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 881.133662] env[61868]: DEBUG oslo_vmware.api [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]521ce4c2-057e-d00d-2f02-4f09bb0f5665, 'name': SearchDatastore_Task, 'duration_secs': 0.015497} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.133914] env[61868]: DEBUG oslo_concurrency.lockutils [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.134226] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] b7055e5a-37d2-42d5-bab0-1e70faf52b58/b7055e5a-37d2-42d5-bab0-1e70faf52b58.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 881.134485] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3e5034f4-8b61-49e6-82aa-2f9e7241890c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.140667] env[61868]: DEBUG oslo_vmware.api [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 881.140667] env[61868]: value = "task-1315558" [ 881.140667] env[61868]: _type = "Task" [ 881.140667] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.149118] env[61868]: DEBUG oslo_vmware.api [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315558, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.381748] env[61868]: DEBUG nova.network.neutron [req-35c48f24-7bbf-496d-9486-2481f2133676 req-05989015-6cfa-44a0-9865-8829f2f05c62 service nova] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Updated VIF entry in instance network info cache for port 981ca2e3-c2d7-4d28-bda6-68ddd6d57bc5. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 881.381748] env[61868]: DEBUG nova.network.neutron [req-35c48f24-7bbf-496d-9486-2481f2133676 req-05989015-6cfa-44a0-9865-8829f2f05c62 service nova] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Updating instance_info_cache with network_info: [{"id": "981ca2e3-c2d7-4d28-bda6-68ddd6d57bc5", "address": "fa:16:3e:d8:2f:78", "network": {"id": "e3a6a163-518e-43ca-b17d-d38d64a3d2a4", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1496913280-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "50658d903dee454eb544ebf92621faa2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24210a23-d8ac-4f4f-84ac-dc0636de9a72", "external-id": "nsx-vlan-transportzone-257", "segmentation_id": 257, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap981ca2e3-c2", "ovs_interfaceid": "981ca2e3-c2d7-4d28-bda6-68ddd6d57bc5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.442962] env[61868]: DEBUG oslo_concurrency.lockutils [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "4e9d75e8-ca33-4e33-baa3-cb7d596f584e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.443290] env[61868]: DEBUG oslo_concurrency.lockutils [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "4e9d75e8-ca33-4e33-baa3-cb7d596f584e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.447561] env[61868]: DEBUG nova.scheduler.client.report [None req-0a2a4111-8de4-4541-8ce5-040de2bcae6b tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 881.528401] env[61868]: DEBUG oslo_vmware.api [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Task: {'id': task-1315557, 'name': Rename_Task, 'duration_secs': 0.350567} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.531022] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 881.531022] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-83e85763-bb29-41e6-ac10-fcdccd1c88c7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.541021] env[61868]: DEBUG oslo_vmware.api [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Waiting for the task: (returnval){ [ 881.541021] env[61868]: value = "task-1315559" [ 881.541021] env[61868]: _type = "Task" [ 881.541021] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.554227] env[61868]: DEBUG oslo_vmware.api [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Task: {'id': task-1315559, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.584467] env[61868]: DEBUG oslo_concurrency.lockutils [None req-be680e47-2395-47e2-bfb6-da2956f42022 tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Acquiring lock "refresh_cache-c6108cf2-b597-4ca7-8f57-12ea82cc6c2c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.595182] env[61868]: DEBUG nova.compute.manager [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 881.650820] env[61868]: DEBUG oslo_vmware.api [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315558, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.885466] env[61868]: DEBUG oslo_concurrency.lockutils [req-35c48f24-7bbf-496d-9486-2481f2133676 req-05989015-6cfa-44a0-9865-8829f2f05c62 service nova] Releasing lock "refresh_cache-c6108cf2-b597-4ca7-8f57-12ea82cc6c2c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.886346] env[61868]: DEBUG oslo_concurrency.lockutils [None req-be680e47-2395-47e2-bfb6-da2956f42022 tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Acquired lock "refresh_cache-c6108cf2-b597-4ca7-8f57-12ea82cc6c2c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.886346] env[61868]: DEBUG nova.network.neutron [None req-be680e47-2395-47e2-bfb6-da2956f42022 tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 881.955292] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0a2a4111-8de4-4541-8ce5-040de2bcae6b tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.950s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.957929] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.898s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.959475] env[61868]: INFO nova.compute.claims [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 881.991316] env[61868]: INFO nova.scheduler.client.report [None req-0a2a4111-8de4-4541-8ce5-040de2bcae6b tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Deleted allocations for instance ca85dfd1-e794-41dc-a337-fbd99998e1fa [ 882.031343] env[61868]: DEBUG nova.compute.manager [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 882.046674] env[61868]: DEBUG nova.compute.manager [req-678a92fd-ffd3-4fe4-8331-b697bc09f5b4 req-ea6dbd7c-4a6b-4ab5-802d-233ce7e6113d service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Received event network-changed-9771cedf-dae3-41e8-ad4b-aec421274d6c {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 882.046862] env[61868]: DEBUG nova.compute.manager [req-678a92fd-ffd3-4fe4-8331-b697bc09f5b4 req-ea6dbd7c-4a6b-4ab5-802d-233ce7e6113d service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Refreshing instance network info cache due to event network-changed-9771cedf-dae3-41e8-ad4b-aec421274d6c. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 882.049289] env[61868]: DEBUG oslo_concurrency.lockutils [req-678a92fd-ffd3-4fe4-8331-b697bc09f5b4 req-ea6dbd7c-4a6b-4ab5-802d-233ce7e6113d service nova] Acquiring lock "refresh_cache-5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.049289] env[61868]: DEBUG oslo_concurrency.lockutils [req-678a92fd-ffd3-4fe4-8331-b697bc09f5b4 req-ea6dbd7c-4a6b-4ab5-802d-233ce7e6113d service nova] Acquired lock "refresh_cache-5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.049289] env[61868]: DEBUG nova.network.neutron [req-678a92fd-ffd3-4fe4-8331-b697bc09f5b4 req-ea6dbd7c-4a6b-4ab5-802d-233ce7e6113d service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Refreshing network info cache for port 9771cedf-dae3-41e8-ad4b-aec421274d6c {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 882.055028] env[61868]: DEBUG oslo_vmware.api [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Task: {'id': task-1315559, 'name': PowerOnVM_Task} progress is 78%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.063379] env[61868]: DEBUG nova.virt.hardware [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 882.063613] env[61868]: DEBUG nova.virt.hardware [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 882.063770] env[61868]: DEBUG nova.virt.hardware [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 882.063957] env[61868]: DEBUG nova.virt.hardware [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 882.064130] env[61868]: DEBUG nova.virt.hardware [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 882.064330] env[61868]: DEBUG nova.virt.hardware [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 882.064485] env[61868]: DEBUG nova.virt.hardware [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 882.064643] env[61868]: DEBUG nova.virt.hardware [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 882.064811] env[61868]: DEBUG nova.virt.hardware [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 882.064981] env[61868]: DEBUG nova.virt.hardware [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 882.065169] env[61868]: DEBUG nova.virt.hardware [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 882.066184] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a56d2fd-3ea9-469c-abbc-6935c30854e4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.076596] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac918981-67f6-4011-bc14-d1c6c70ea236 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.115801] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.151749] env[61868]: DEBUG oslo_vmware.api [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315558, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.516522} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.152034] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] b7055e5a-37d2-42d5-bab0-1e70faf52b58/b7055e5a-37d2-42d5-bab0-1e70faf52b58.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 882.152256] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 882.152507] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c08d5be3-fabd-421b-a7dc-d68fbc6aba5f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.158493] env[61868]: DEBUG oslo_vmware.api [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 882.158493] env[61868]: value = "task-1315560" [ 882.158493] env[61868]: _type = "Task" [ 882.158493] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.165820] env[61868]: DEBUG oslo_vmware.api [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315560, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.499597] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0a2a4111-8de4-4541-8ce5-040de2bcae6b tempest-ServerPasswordTestJSON-748979559 tempest-ServerPasswordTestJSON-748979559-project-member] Lock "ca85dfd1-e794-41dc-a337-fbd99998e1fa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.869s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.553588] env[61868]: DEBUG oslo_vmware.api [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Task: {'id': task-1315559, 'name': PowerOnVM_Task, 'duration_secs': 0.823691} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.554269] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 882.554512] env[61868]: INFO nova.compute.manager [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Took 7.88 seconds to spawn the instance on the hypervisor. [ 882.554743] env[61868]: DEBUG nova.compute.manager [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 882.556105] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-293b9f5e-4e5b-48fb-9782-f25b3dd192f2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.668085] env[61868]: DEBUG oslo_vmware.api [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315560, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065051} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.668721] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 882.669166] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a730c9d1-5afa-4697-b9c3-70055d585b85 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.698602] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] b7055e5a-37d2-42d5-bab0-1e70faf52b58/b7055e5a-37d2-42d5-bab0-1e70faf52b58.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 882.698932] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8af225ed-e360-476b-92ea-e763bbb9186e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.728804] env[61868]: DEBUG oslo_vmware.api [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 882.728804] env[61868]: value = "task-1315561" [ 882.728804] env[61868]: _type = "Task" [ 882.728804] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.738495] env[61868]: DEBUG oslo_vmware.api [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315561, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.814813] env[61868]: DEBUG nova.network.neutron [None req-be680e47-2395-47e2-bfb6-da2956f42022 tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Updating instance_info_cache with network_info: [{"id": "981ca2e3-c2d7-4d28-bda6-68ddd6d57bc5", "address": "fa:16:3e:d8:2f:78", "network": {"id": "e3a6a163-518e-43ca-b17d-d38d64a3d2a4", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1496913280-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "50658d903dee454eb544ebf92621faa2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24210a23-d8ac-4f4f-84ac-dc0636de9a72", "external-id": "nsx-vlan-transportzone-257", "segmentation_id": 257, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap981ca2e3-c2", "ovs_interfaceid": "981ca2e3-c2d7-4d28-bda6-68ddd6d57bc5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.886131] env[61868]: DEBUG nova.network.neutron [req-678a92fd-ffd3-4fe4-8331-b697bc09f5b4 req-ea6dbd7c-4a6b-4ab5-802d-233ce7e6113d service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Updated VIF entry in instance network info cache for port 9771cedf-dae3-41e8-ad4b-aec421274d6c. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 882.886307] env[61868]: DEBUG nova.network.neutron [req-678a92fd-ffd3-4fe4-8331-b697bc09f5b4 req-ea6dbd7c-4a6b-4ab5-802d-233ce7e6113d service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Updating instance_info_cache with network_info: [{"id": "9771cedf-dae3-41e8-ad4b-aec421274d6c", "address": "fa:16:3e:4f:5f:b8", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.244", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9771cedf-da", "ovs_interfaceid": "9771cedf-dae3-41e8-ad4b-aec421274d6c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.078329] env[61868]: INFO nova.compute.manager [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Took 25.87 seconds to build instance. [ 883.124732] env[61868]: DEBUG nova.network.neutron [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Successfully updated port: 29e8f7bd-db77-4071-b4c6-79e8338c2e82 {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 883.243550] env[61868]: DEBUG oslo_vmware.api [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315561, 'name': ReconfigVM_Task, 'duration_secs': 0.273944} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.243550] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Reconfigured VM instance instance-00000049 to attach disk [datastore1] b7055e5a-37d2-42d5-bab0-1e70faf52b58/b7055e5a-37d2-42d5-bab0-1e70faf52b58.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 883.244054] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e0590bea-a7cf-4709-a3b3-1f5bce90ee42 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.253062] env[61868]: DEBUG oslo_vmware.api [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 883.253062] env[61868]: value = "task-1315562" [ 883.253062] env[61868]: _type = "Task" [ 883.253062] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.269084] env[61868]: DEBUG oslo_vmware.api [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315562, 'name': Rename_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.298440] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12592138-bc37-41a7-8b42-b0f704233ca3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.307260] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a182b75-5e85-41a5-a87d-34a923cfbc17 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.345372] env[61868]: DEBUG oslo_concurrency.lockutils [None req-be680e47-2395-47e2-bfb6-da2956f42022 tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Releasing lock "refresh_cache-c6108cf2-b597-4ca7-8f57-12ea82cc6c2c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.347960] env[61868]: DEBUG nova.compute.manager [None req-be680e47-2395-47e2-bfb6-da2956f42022 tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 883.348858] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d3d8dc8-675d-464e-8091-e85ebf888631 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.354048] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eaf092f-9e9d-4a6b-841b-8d2803e58d93 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.369308] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69d4d93f-14cb-482b-8610-7d06fb34b3b8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.392021] env[61868]: DEBUG oslo_concurrency.lockutils [req-678a92fd-ffd3-4fe4-8331-b697bc09f5b4 req-ea6dbd7c-4a6b-4ab5-802d-233ce7e6113d service nova] Releasing lock "refresh_cache-5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.392021] env[61868]: DEBUG nova.compute.provider_tree [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 883.585687] env[61868]: DEBUG oslo_concurrency.lockutils [None req-993dcb6a-96c6-49d6-bf46-27a8f91bfad9 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Lock "b9b5be37-6b30-4229-9c8d-3ee9d30db119" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 67.228s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.629751] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Acquiring lock "refresh_cache-f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.630042] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Acquired lock "refresh_cache-f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.630042] env[61868]: DEBUG nova.network.neutron [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 883.764054] env[61868]: DEBUG oslo_vmware.api [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315562, 'name': Rename_Task, 'duration_secs': 0.150013} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.764307] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 883.764953] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5c7b12a8-d17f-4aa4-a8db-36ff685acbd7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.772439] env[61868]: DEBUG oslo_vmware.api [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 883.772439] env[61868]: value = "task-1315563" [ 883.772439] env[61868]: _type = "Task" [ 883.772439] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.781571] env[61868]: DEBUG oslo_vmware.api [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315563, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.894373] env[61868]: DEBUG nova.scheduler.client.report [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 884.085342] env[61868]: DEBUG nova.compute.manager [req-a03c0647-a526-43f9-8346-8b3c345347c1 req-38b83b3e-22d7-49c3-8195-ec4575b73e9c service nova] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Received event network-vif-plugged-29e8f7bd-db77-4071-b4c6-79e8338c2e82 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 884.085540] env[61868]: DEBUG oslo_concurrency.lockutils [req-a03c0647-a526-43f9-8346-8b3c345347c1 req-38b83b3e-22d7-49c3-8195-ec4575b73e9c service nova] Acquiring lock "f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.085751] env[61868]: DEBUG oslo_concurrency.lockutils [req-a03c0647-a526-43f9-8346-8b3c345347c1 req-38b83b3e-22d7-49c3-8195-ec4575b73e9c service nova] Lock "f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.085970] env[61868]: DEBUG oslo_concurrency.lockutils [req-a03c0647-a526-43f9-8346-8b3c345347c1 req-38b83b3e-22d7-49c3-8195-ec4575b73e9c service nova] Lock "f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.086169] env[61868]: DEBUG nova.compute.manager [req-a03c0647-a526-43f9-8346-8b3c345347c1 req-38b83b3e-22d7-49c3-8195-ec4575b73e9c service nova] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] No waiting events found dispatching network-vif-plugged-29e8f7bd-db77-4071-b4c6-79e8338c2e82 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 884.086353] env[61868]: WARNING nova.compute.manager [req-a03c0647-a526-43f9-8346-8b3c345347c1 req-38b83b3e-22d7-49c3-8195-ec4575b73e9c service nova] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Received unexpected event network-vif-plugged-29e8f7bd-db77-4071-b4c6-79e8338c2e82 for instance with vm_state building and task_state spawning. [ 884.086522] env[61868]: DEBUG nova.compute.manager [req-a03c0647-a526-43f9-8346-8b3c345347c1 req-38b83b3e-22d7-49c3-8195-ec4575b73e9c service nova] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Received event network-changed-29e8f7bd-db77-4071-b4c6-79e8338c2e82 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 884.086681] env[61868]: DEBUG nova.compute.manager [req-a03c0647-a526-43f9-8346-8b3c345347c1 req-38b83b3e-22d7-49c3-8195-ec4575b73e9c service nova] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Refreshing instance network info cache due to event network-changed-29e8f7bd-db77-4071-b4c6-79e8338c2e82. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 884.086872] env[61868]: DEBUG oslo_concurrency.lockutils [req-a03c0647-a526-43f9-8346-8b3c345347c1 req-38b83b3e-22d7-49c3-8195-ec4575b73e9c service nova] Acquiring lock "refresh_cache-f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.088565] env[61868]: DEBUG nova.compute.manager [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 884.163297] env[61868]: DEBUG nova.network.neutron [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 884.283985] env[61868]: DEBUG oslo_vmware.api [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315563, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.286122] env[61868]: DEBUG nova.network.neutron [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Updating instance_info_cache with network_info: [{"id": "29e8f7bd-db77-4071-b4c6-79e8338c2e82", "address": "fa:16:3e:be:13:44", "network": {"id": "4befda35-5c88-4ac7-b8ef-128ddae97a7d", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-113207814-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "78a206c3daa14cc58e7fe8546f565ae0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd72ef32-a57c-43b0-93df-e8a030987d44", "external-id": "nsx-vlan-transportzone-340", "segmentation_id": 340, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap29e8f7bd-db", "ovs_interfaceid": "29e8f7bd-db77-4071-b4c6-79e8338c2e82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.380196] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c49d4c86-ec6e-4240-a934-8ba1e0888c44 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.391796] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-be680e47-2395-47e2-bfb6-da2956f42022 tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Doing hard reboot of VM {{(pid=61868) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 884.391796] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-5216284b-0ed5-42e9-a513-7a97120d5bd1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.399559] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.442s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.400227] env[61868]: DEBUG nova.compute.manager [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 884.404977] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.195s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.405283] env[61868]: DEBUG nova.objects.instance [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Lazy-loading 'resources' on Instance uuid d35953d8-692d-498e-baf5-96ef381ce12b {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 884.407266] env[61868]: DEBUG oslo_vmware.api [None req-be680e47-2395-47e2-bfb6-da2956f42022 tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Waiting for the task: (returnval){ [ 884.407266] env[61868]: value = "task-1315564" [ 884.407266] env[61868]: _type = "Task" [ 884.407266] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.427017] env[61868]: DEBUG oslo_vmware.api [None req-be680e47-2395-47e2-bfb6-da2956f42022 tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Task: {'id': task-1315564, 'name': ResetVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.615244] env[61868]: DEBUG oslo_concurrency.lockutils [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.785382] env[61868]: DEBUG oslo_vmware.api [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315563, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.789092] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Releasing lock "refresh_cache-f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.789533] env[61868]: DEBUG nova.compute.manager [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Instance network_info: |[{"id": "29e8f7bd-db77-4071-b4c6-79e8338c2e82", "address": "fa:16:3e:be:13:44", "network": {"id": "4befda35-5c88-4ac7-b8ef-128ddae97a7d", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-113207814-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "78a206c3daa14cc58e7fe8546f565ae0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd72ef32-a57c-43b0-93df-e8a030987d44", "external-id": "nsx-vlan-transportzone-340", "segmentation_id": 340, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap29e8f7bd-db", "ovs_interfaceid": "29e8f7bd-db77-4071-b4c6-79e8338c2e82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 884.789812] env[61868]: DEBUG oslo_concurrency.lockutils [req-a03c0647-a526-43f9-8346-8b3c345347c1 req-38b83b3e-22d7-49c3-8195-ec4575b73e9c service nova] Acquired lock "refresh_cache-f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.790036] env[61868]: DEBUG nova.network.neutron [req-a03c0647-a526-43f9-8346-8b3c345347c1 req-38b83b3e-22d7-49c3-8195-ec4575b73e9c service nova] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Refreshing network info cache for port 29e8f7bd-db77-4071-b4c6-79e8338c2e82 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 884.791462] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:be:13:44', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dd72ef32-a57c-43b0-93df-e8a030987d44', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '29e8f7bd-db77-4071-b4c6-79e8338c2e82', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 884.799762] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Creating folder: Project (78a206c3daa14cc58e7fe8546f565ae0). Parent ref: group-v281478. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 884.803022] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-23279c22-c114-44ee-96e8-86017f7a0238 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.817452] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Created folder: Project (78a206c3daa14cc58e7fe8546f565ae0) in parent group-v281478. [ 884.817452] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Creating folder: Instances. Parent ref: group-v281554. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 884.817452] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1c6982ce-8883-42be-b63f-a330d50a0f72 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.828194] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Created folder: Instances in parent group-v281554. [ 884.828471] env[61868]: DEBUG oslo.service.loopingcall [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 884.828671] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 884.828878] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-709d1e06-0a83-4fe7-8167-a77fa6c10f41 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.853059] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 884.853059] env[61868]: value = "task-1315567" [ 884.853059] env[61868]: _type = "Task" [ 884.853059] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.862474] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315567, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.911291] env[61868]: DEBUG nova.compute.utils [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 884.916126] env[61868]: DEBUG nova.compute.manager [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 884.916800] env[61868]: DEBUG nova.network.neutron [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 884.929605] env[61868]: DEBUG oslo_vmware.api [None req-be680e47-2395-47e2-bfb6-da2956f42022 tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Task: {'id': task-1315564, 'name': ResetVM_Task, 'duration_secs': 0.11358} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.929918] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-be680e47-2395-47e2-bfb6-da2956f42022 tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Did hard reboot of VM {{(pid=61868) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 884.930174] env[61868]: DEBUG nova.compute.manager [None req-be680e47-2395-47e2-bfb6-da2956f42022 tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 884.931048] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ce9dd0a-3efb-441b-bd11-418157b8bd5f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.981851] env[61868]: DEBUG nova.policy [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0b615dd3d1064706bea31b52fbfa9a61', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5658848e1b0e42929c04a4a8de40a291', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 885.068165] env[61868]: DEBUG nova.network.neutron [req-a03c0647-a526-43f9-8346-8b3c345347c1 req-38b83b3e-22d7-49c3-8195-ec4575b73e9c service nova] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Updated VIF entry in instance network info cache for port 29e8f7bd-db77-4071-b4c6-79e8338c2e82. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 885.071388] env[61868]: DEBUG nova.network.neutron [req-a03c0647-a526-43f9-8346-8b3c345347c1 req-38b83b3e-22d7-49c3-8195-ec4575b73e9c service nova] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Updating instance_info_cache with network_info: [{"id": "29e8f7bd-db77-4071-b4c6-79e8338c2e82", "address": "fa:16:3e:be:13:44", "network": {"id": "4befda35-5c88-4ac7-b8ef-128ddae97a7d", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-113207814-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "78a206c3daa14cc58e7fe8546f565ae0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd72ef32-a57c-43b0-93df-e8a030987d44", "external-id": "nsx-vlan-transportzone-340", "segmentation_id": 340, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap29e8f7bd-db", "ovs_interfaceid": "29e8f7bd-db77-4071-b4c6-79e8338c2e82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.286456] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c2232f1-8d74-4853-8d6d-806a8db3b833 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.293122] env[61868]: DEBUG oslo_vmware.api [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315563, 'name': PowerOnVM_Task, 'duration_secs': 1.017337} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.294113] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 885.294437] env[61868]: INFO nova.compute.manager [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Took 8.14 seconds to spawn the instance on the hypervisor. [ 885.294696] env[61868]: DEBUG nova.compute.manager [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 885.295911] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddc5db3c-40c0-4781-acd5-0ab8268ef884 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.303228] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa95b052-f809-4abe-83da-410d749ba671 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.348026] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d544f718-2baa-44de-a7eb-431a7fa7275f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.357766] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-131bec58-211a-4db5-8444-f29ae039d66a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.368414] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315567, 'name': CreateVM_Task, 'duration_secs': 0.434153} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.376770] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 885.377433] env[61868]: DEBUG nova.compute.provider_tree [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 885.379200] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.379361] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.379690] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 885.380156] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b9a1e9f5-fb3a-4271-831b-4dab050bc54a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.386303] env[61868]: DEBUG oslo_vmware.api [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Waiting for the task: (returnval){ [ 885.386303] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5248808a-4e7c-77bf-6c7c-f7c628f91cfb" [ 885.386303] env[61868]: _type = "Task" [ 885.386303] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.395509] env[61868]: DEBUG oslo_vmware.api [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5248808a-4e7c-77bf-6c7c-f7c628f91cfb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.417209] env[61868]: DEBUG nova.compute.manager [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 885.452329] env[61868]: DEBUG oslo_concurrency.lockutils [None req-be680e47-2395-47e2-bfb6-da2956f42022 tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Lock "c6108cf2-b597-4ca7-8f57-12ea82cc6c2c" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.394s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.573794] env[61868]: DEBUG oslo_concurrency.lockutils [req-a03c0647-a526-43f9-8346-8b3c345347c1 req-38b83b3e-22d7-49c3-8195-ec4575b73e9c service nova] Releasing lock "refresh_cache-f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.618593] env[61868]: DEBUG nova.network.neutron [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Successfully created port: 4186aca7-63b5-4b24-ab7c-de8f87064242 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 885.823537] env[61868]: INFO nova.compute.manager [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Took 27.90 seconds to build instance. [ 885.881195] env[61868]: DEBUG nova.scheduler.client.report [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 885.897495] env[61868]: DEBUG oslo_vmware.api [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5248808a-4e7c-77bf-6c7c-f7c628f91cfb, 'name': SearchDatastore_Task, 'duration_secs': 0.016838} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.897811] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.898063] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 885.898318] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.898445] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.898628] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 885.898889] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-be50fc30-935b-4192-b556-923aea5a4973 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.912750] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 885.912915] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 885.913634] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f922d100-0d5f-4c4e-9086-e7744c6b49db {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.920045] env[61868]: DEBUG oslo_vmware.api [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Waiting for the task: (returnval){ [ 885.920045] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52c21939-c65b-ac94-76a6-8257996e4c9e" [ 885.920045] env[61868]: _type = "Task" [ 885.920045] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.932313] env[61868]: DEBUG oslo_vmware.api [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52c21939-c65b-ac94-76a6-8257996e4c9e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.326461] env[61868]: DEBUG oslo_concurrency.lockutils [None req-64addc39-2061-4383-8bb7-358950fbe217 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "b7055e5a-37d2-42d5-bab0-1e70faf52b58" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 65.312s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.391017] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.984s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.392433] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.820s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.421111] env[61868]: INFO nova.scheduler.client.report [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Deleted allocations for instance d35953d8-692d-498e-baf5-96ef381ce12b [ 886.431167] env[61868]: DEBUG nova.compute.manager [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 886.441542] env[61868]: DEBUG oslo_vmware.api [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52c21939-c65b-ac94-76a6-8257996e4c9e, 'name': SearchDatastore_Task, 'duration_secs': 0.022201} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.442581] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eeb8598a-89ed-4655-8949-8397e7dba020 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.453050] env[61868]: DEBUG oslo_vmware.api [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Waiting for the task: (returnval){ [ 886.453050] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]520310e7-b78a-060d-674c-819c44e2df1a" [ 886.453050] env[61868]: _type = "Task" [ 886.453050] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.453478] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0f7f9081-032d-4547-939e-b812c9bdc0c7 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Acquiring lock "b9b5be37-6b30-4229-9c8d-3ee9d30db119" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.454178] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0f7f9081-032d-4547-939e-b812c9bdc0c7 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Lock "b9b5be37-6b30-4229-9c8d-3ee9d30db119" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.454437] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0f7f9081-032d-4547-939e-b812c9bdc0c7 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Acquiring lock "b9b5be37-6b30-4229-9c8d-3ee9d30db119-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.454879] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0f7f9081-032d-4547-939e-b812c9bdc0c7 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Lock "b9b5be37-6b30-4229-9c8d-3ee9d30db119-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.455118] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0f7f9081-032d-4547-939e-b812c9bdc0c7 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Lock "b9b5be37-6b30-4229-9c8d-3ee9d30db119-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.457529] env[61868]: INFO nova.compute.manager [None req-0f7f9081-032d-4547-939e-b812c9bdc0c7 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Terminating instance [ 886.460147] env[61868]: DEBUG nova.compute.manager [None req-0f7f9081-032d-4547-939e-b812c9bdc0c7 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 886.460919] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0f7f9081-032d-4547-939e-b812c9bdc0c7 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 886.465130] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6bba286-b399-41d2-ad7a-292e9ce172c4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.482157] env[61868]: DEBUG oslo_vmware.api [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]520310e7-b78a-060d-674c-819c44e2df1a, 'name': SearchDatastore_Task, 'duration_secs': 0.011971} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.486897] env[61868]: DEBUG nova.virt.hardware [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 886.487801] env[61868]: DEBUG nova.virt.hardware [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 886.487801] env[61868]: DEBUG nova.virt.hardware [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 886.487801] env[61868]: DEBUG nova.virt.hardware [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 886.487801] env[61868]: DEBUG nova.virt.hardware [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 886.487801] env[61868]: DEBUG nova.virt.hardware [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 886.488158] env[61868]: DEBUG nova.virt.hardware [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 886.488158] env[61868]: DEBUG nova.virt.hardware [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 886.488265] env[61868]: DEBUG nova.virt.hardware [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 886.490090] env[61868]: DEBUG nova.virt.hardware [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 886.490090] env[61868]: DEBUG nova.virt.hardware [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 886.490090] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.490090] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3/f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 886.491206] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f7f9081-032d-4547-939e-b812c9bdc0c7 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 886.491206] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13ea8e6c-cf9a-4955-92cc-b040fd7283fc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.496367] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-742d0586-b3f7-480d-9625-2a355b75f58d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.498624] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d92a8f1f-0cc1-4b59-82e7-f5a130f5b7a1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.508698] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf95b664-a575-4f03-8120-9c6d242e7a94 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.514441] env[61868]: DEBUG oslo_vmware.api [None req-0f7f9081-032d-4547-939e-b812c9bdc0c7 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Waiting for the task: (returnval){ [ 886.514441] env[61868]: value = "task-1315569" [ 886.514441] env[61868]: _type = "Task" [ 886.514441] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.514805] env[61868]: DEBUG oslo_vmware.api [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Waiting for the task: (returnval){ [ 886.514805] env[61868]: value = "task-1315568" [ 886.514805] env[61868]: _type = "Task" [ 886.514805] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.538915] env[61868]: DEBUG oslo_vmware.api [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': task-1315568, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.542554] env[61868]: DEBUG oslo_vmware.api [None req-0f7f9081-032d-4547-939e-b812c9bdc0c7 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Task: {'id': task-1315569, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.660957] env[61868]: DEBUG nova.compute.manager [req-e555f7e2-c955-4bf1-ae4b-e2ff4fbf9353 req-874bba43-4801-4923-9762-c1d8e1ec7866 service nova] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Received event network-changed-981ca2e3-c2d7-4d28-bda6-68ddd6d57bc5 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 886.661387] env[61868]: DEBUG nova.compute.manager [req-e555f7e2-c955-4bf1-ae4b-e2ff4fbf9353 req-874bba43-4801-4923-9762-c1d8e1ec7866 service nova] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Refreshing instance network info cache due to event network-changed-981ca2e3-c2d7-4d28-bda6-68ddd6d57bc5. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 886.661387] env[61868]: DEBUG oslo_concurrency.lockutils [req-e555f7e2-c955-4bf1-ae4b-e2ff4fbf9353 req-874bba43-4801-4923-9762-c1d8e1ec7866 service nova] Acquiring lock "refresh_cache-c6108cf2-b597-4ca7-8f57-12ea82cc6c2c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.661958] env[61868]: DEBUG oslo_concurrency.lockutils [req-e555f7e2-c955-4bf1-ae4b-e2ff4fbf9353 req-874bba43-4801-4923-9762-c1d8e1ec7866 service nova] Acquired lock "refresh_cache-c6108cf2-b597-4ca7-8f57-12ea82cc6c2c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.661958] env[61868]: DEBUG nova.network.neutron [req-e555f7e2-c955-4bf1-ae4b-e2ff4fbf9353 req-874bba43-4801-4923-9762-c1d8e1ec7866 service nova] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Refreshing network info cache for port 981ca2e3-c2d7-4d28-bda6-68ddd6d57bc5 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 886.779395] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6d5ed2d-e4c5-49d5-ac68-9b719cc7fa2c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.789072] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af63195f-3ad0-4009-a362-cb93eac0c682 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.828221] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46123c0e-bdc8-4b99-8359-ee6178923a60 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.839860] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59a4f2f2-4f4b-415c-aa21-e6b6a6093c35 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.858378] env[61868]: DEBUG nova.compute.provider_tree [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 886.931584] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6ead97a6-1afc-49f8-9845-280ae325b079 tempest-ServerShowV247Test-605785174 tempest-ServerShowV247Test-605785174-project-member] Lock "d35953d8-692d-498e-baf5-96ef381ce12b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.436s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.041872] env[61868]: DEBUG oslo_vmware.api [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': task-1315568, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.042698] env[61868]: DEBUG oslo_vmware.api [None req-0f7f9081-032d-4547-939e-b812c9bdc0c7 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Task: {'id': task-1315569, 'name': PowerOffVM_Task, 'duration_secs': 0.226533} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.042959] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f7f9081-032d-4547-939e-b812c9bdc0c7 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 887.043138] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0f7f9081-032d-4547-939e-b812c9bdc0c7 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 887.043403] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9f39dd2c-1a80-4633-a3a3-0cb92cb8aa45 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.133600] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0f7f9081-032d-4547-939e-b812c9bdc0c7 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 887.133872] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0f7f9081-032d-4547-939e-b812c9bdc0c7 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 887.134120] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f7f9081-032d-4547-939e-b812c9bdc0c7 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Deleting the datastore file [datastore1] b9b5be37-6b30-4229-9c8d-3ee9d30db119 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 887.134440] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6312b308-81bb-4fac-bf5e-6dac72e34946 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.144222] env[61868]: DEBUG oslo_vmware.api [None req-0f7f9081-032d-4547-939e-b812c9bdc0c7 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Waiting for the task: (returnval){ [ 887.144222] env[61868]: value = "task-1315571" [ 887.144222] env[61868]: _type = "Task" [ 887.144222] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.161218] env[61868]: DEBUG oslo_vmware.api [None req-0f7f9081-032d-4547-939e-b812c9bdc0c7 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Task: {'id': task-1315571, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.329223] env[61868]: DEBUG oslo_concurrency.lockutils [None req-805d5425-8b50-412b-81b4-c3b60c28ab13 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "b7055e5a-37d2-42d5-bab0-1e70faf52b58" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.329519] env[61868]: DEBUG oslo_concurrency.lockutils [None req-805d5425-8b50-412b-81b4-c3b60c28ab13 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "b7055e5a-37d2-42d5-bab0-1e70faf52b58" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.329743] env[61868]: DEBUG oslo_concurrency.lockutils [None req-805d5425-8b50-412b-81b4-c3b60c28ab13 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "b7055e5a-37d2-42d5-bab0-1e70faf52b58-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.330306] env[61868]: DEBUG oslo_concurrency.lockutils [None req-805d5425-8b50-412b-81b4-c3b60c28ab13 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "b7055e5a-37d2-42d5-bab0-1e70faf52b58-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.330306] env[61868]: DEBUG oslo_concurrency.lockutils [None req-805d5425-8b50-412b-81b4-c3b60c28ab13 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "b7055e5a-37d2-42d5-bab0-1e70faf52b58-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.334824] env[61868]: INFO nova.compute.manager [None req-805d5425-8b50-412b-81b4-c3b60c28ab13 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Terminating instance [ 887.337111] env[61868]: DEBUG nova.compute.manager [None req-805d5425-8b50-412b-81b4-c3b60c28ab13 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 887.337263] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-805d5425-8b50-412b-81b4-c3b60c28ab13 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 887.338217] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dca9a56-2b3c-42a8-a0ae-8fd35d84c773 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.355856] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-805d5425-8b50-412b-81b4-c3b60c28ab13 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 887.356245] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e599e41c-dc67-4fb3-a26d-0a70c267100a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.362113] env[61868]: DEBUG nova.scheduler.client.report [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 887.370267] env[61868]: DEBUG oslo_vmware.api [None req-805d5425-8b50-412b-81b4-c3b60c28ab13 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 887.370267] env[61868]: value = "task-1315572" [ 887.370267] env[61868]: _type = "Task" [ 887.370267] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.382799] env[61868]: DEBUG oslo_vmware.api [None req-805d5425-8b50-412b-81b4-c3b60c28ab13 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315572, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.491925] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4dcc6a68-7a7e-4a85-8566-794266e676cd tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Acquiring lock "c6108cf2-b597-4ca7-8f57-12ea82cc6c2c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.492239] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4dcc6a68-7a7e-4a85-8566-794266e676cd tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Lock "c6108cf2-b597-4ca7-8f57-12ea82cc6c2c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.493491] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4dcc6a68-7a7e-4a85-8566-794266e676cd tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Acquiring lock "c6108cf2-b597-4ca7-8f57-12ea82cc6c2c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.493491] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4dcc6a68-7a7e-4a85-8566-794266e676cd tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Lock "c6108cf2-b597-4ca7-8f57-12ea82cc6c2c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.493491] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4dcc6a68-7a7e-4a85-8566-794266e676cd tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Lock "c6108cf2-b597-4ca7-8f57-12ea82cc6c2c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.496173] env[61868]: INFO nova.compute.manager [None req-4dcc6a68-7a7e-4a85-8566-794266e676cd tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Terminating instance [ 887.498242] env[61868]: DEBUG nova.compute.manager [None req-4dcc6a68-7a7e-4a85-8566-794266e676cd tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 887.498468] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-4dcc6a68-7a7e-4a85-8566-794266e676cd tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 887.499334] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfaaff57-47f8-4f7a-bcfd-9e83c40a2035 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.509143] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-4dcc6a68-7a7e-4a85-8566-794266e676cd tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 887.509567] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7928115b-f645-45d4-91ed-e10fbf75ea5b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.519819] env[61868]: DEBUG oslo_vmware.api [None req-4dcc6a68-7a7e-4a85-8566-794266e676cd tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Waiting for the task: (returnval){ [ 887.519819] env[61868]: value = "task-1315573" [ 887.519819] env[61868]: _type = "Task" [ 887.519819] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.538726] env[61868]: DEBUG oslo_vmware.api [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': task-1315568, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.737528} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.538982] env[61868]: DEBUG oslo_vmware.api [None req-4dcc6a68-7a7e-4a85-8566-794266e676cd tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Task: {'id': task-1315573, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.539820] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3/f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 887.539820] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 887.539820] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6c665759-e7f9-4e00-a0a6-c359cd22599d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.547570] env[61868]: DEBUG oslo_vmware.api [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Waiting for the task: (returnval){ [ 887.547570] env[61868]: value = "task-1315574" [ 887.547570] env[61868]: _type = "Task" [ 887.547570] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.557809] env[61868]: DEBUG oslo_vmware.api [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': task-1315574, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.581348] env[61868]: DEBUG nova.network.neutron [req-e555f7e2-c955-4bf1-ae4b-e2ff4fbf9353 req-874bba43-4801-4923-9762-c1d8e1ec7866 service nova] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Updated VIF entry in instance network info cache for port 981ca2e3-c2d7-4d28-bda6-68ddd6d57bc5. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 887.581848] env[61868]: DEBUG nova.network.neutron [req-e555f7e2-c955-4bf1-ae4b-e2ff4fbf9353 req-874bba43-4801-4923-9762-c1d8e1ec7866 service nova] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Updating instance_info_cache with network_info: [{"id": "981ca2e3-c2d7-4d28-bda6-68ddd6d57bc5", "address": "fa:16:3e:d8:2f:78", "network": {"id": "e3a6a163-518e-43ca-b17d-d38d64a3d2a4", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1496913280-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "50658d903dee454eb544ebf92621faa2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24210a23-d8ac-4f4f-84ac-dc0636de9a72", "external-id": "nsx-vlan-transportzone-257", "segmentation_id": 257, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap981ca2e3-c2", "ovs_interfaceid": "981ca2e3-c2d7-4d28-bda6-68ddd6d57bc5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.659873] env[61868]: DEBUG oslo_vmware.api [None req-0f7f9081-032d-4547-939e-b812c9bdc0c7 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Task: {'id': task-1315571, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.333424} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.662108] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f7f9081-032d-4547-939e-b812c9bdc0c7 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 887.662108] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0f7f9081-032d-4547-939e-b812c9bdc0c7 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 887.662108] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0f7f9081-032d-4547-939e-b812c9bdc0c7 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 887.662398] env[61868]: INFO nova.compute.manager [None req-0f7f9081-032d-4547-939e-b812c9bdc0c7 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Took 1.20 seconds to destroy the instance on the hypervisor. [ 887.662825] env[61868]: DEBUG oslo.service.loopingcall [None req-0f7f9081-032d-4547-939e-b812c9bdc0c7 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 887.664305] env[61868]: DEBUG nova.compute.manager [-] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 887.664305] env[61868]: DEBUG nova.network.neutron [-] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 887.670198] env[61868]: DEBUG nova.network.neutron [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Successfully updated port: 4186aca7-63b5-4b24-ab7c-de8f87064242 {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 887.871235] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.479s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.871471] env[61868]: INFO nova.compute.manager [None req-a6e080b2-9f59-4c9f-9413-9310bc59743a tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Successfully reverted task state from rebuilding on failure for instance. [ 887.878554] env[61868]: DEBUG oslo_concurrency.lockutils [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 17.884s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.878794] env[61868]: DEBUG nova.objects.instance [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61868) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 887.891809] env[61868]: DEBUG oslo_vmware.api [None req-805d5425-8b50-412b-81b4-c3b60c28ab13 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315572, 'name': PowerOffVM_Task, 'duration_secs': 0.284142} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.892192] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-805d5425-8b50-412b-81b4-c3b60c28ab13 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 887.892838] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-805d5425-8b50-412b-81b4-c3b60c28ab13 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 887.892838] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-66bbca70-f34d-45b1-91b7-f4b3a74393a3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.039959] env[61868]: DEBUG oslo_vmware.api [None req-4dcc6a68-7a7e-4a85-8566-794266e676cd tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Task: {'id': task-1315573, 'name': PowerOffVM_Task, 'duration_secs': 0.31868} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.041093] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-4dcc6a68-7a7e-4a85-8566-794266e676cd tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 888.041093] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-4dcc6a68-7a7e-4a85-8566-794266e676cd tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 888.041093] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eed1337f-a1a7-423b-a741-768757ae32a0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.060774] env[61868]: DEBUG oslo_vmware.api [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': task-1315574, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08838} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.061160] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 888.062054] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b7f6dd7-24dc-46ff-8ad3-352ecb88425b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.091874] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3/f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 888.092637] env[61868]: DEBUG oslo_concurrency.lockutils [req-e555f7e2-c955-4bf1-ae4b-e2ff4fbf9353 req-874bba43-4801-4923-9762-c1d8e1ec7866 service nova] Releasing lock "refresh_cache-c6108cf2-b597-4ca7-8f57-12ea82cc6c2c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.093135] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c65cafe3-dde6-4959-8b13-9d74c49262bc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.119346] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-4dcc6a68-7a7e-4a85-8566-794266e676cd tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 888.119526] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-4dcc6a68-7a7e-4a85-8566-794266e676cd tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Deleting contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 888.119669] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-4dcc6a68-7a7e-4a85-8566-794266e676cd tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Deleting the datastore file [datastore2] c6108cf2-b597-4ca7-8f57-12ea82cc6c2c {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 888.123984] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9fc5041f-feab-46c2-ae5a-43e53fb45aec {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.127460] env[61868]: DEBUG oslo_vmware.api [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Waiting for the task: (returnval){ [ 888.127460] env[61868]: value = "task-1315577" [ 888.127460] env[61868]: _type = "Task" [ 888.127460] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.137816] env[61868]: DEBUG oslo_vmware.api [None req-4dcc6a68-7a7e-4a85-8566-794266e676cd tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Waiting for the task: (returnval){ [ 888.137816] env[61868]: value = "task-1315578" [ 888.137816] env[61868]: _type = "Task" [ 888.137816] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.146289] env[61868]: DEBUG oslo_vmware.api [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': task-1315577, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.158033] env[61868]: DEBUG oslo_vmware.api [None req-4dcc6a68-7a7e-4a85-8566-794266e676cd tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Task: {'id': task-1315578, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.175815] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "refresh_cache-306f2dc4-933c-4948-8330-eda93931c4e0" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.175815] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquired lock "refresh_cache-306f2dc4-933c-4948-8330-eda93931c4e0" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.175815] env[61868]: DEBUG nova.network.neutron [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 888.257479] env[61868]: DEBUG nova.compute.manager [req-8fbc765a-a49d-43d6-81f7-e10d67e9666d req-5d1c4e26-33a3-43f5-8750-a8866f52edda service nova] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Received event network-vif-deleted-ae9b8f9e-17ec-4c1f-8052-8467a711410c {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 888.257674] env[61868]: INFO nova.compute.manager [req-8fbc765a-a49d-43d6-81f7-e10d67e9666d req-5d1c4e26-33a3-43f5-8750-a8866f52edda service nova] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Neutron deleted interface ae9b8f9e-17ec-4c1f-8052-8467a711410c; detaching it from the instance and deleting it from the info cache [ 888.257846] env[61868]: DEBUG nova.network.neutron [req-8fbc765a-a49d-43d6-81f7-e10d67e9666d req-5d1c4e26-33a3-43f5-8750-a8866f52edda service nova] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.639253] env[61868]: DEBUG oslo_vmware.api [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': task-1315577, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.648681] env[61868]: DEBUG oslo_vmware.api [None req-4dcc6a68-7a7e-4a85-8566-794266e676cd tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Task: {'id': task-1315578, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.249623} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.648681] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-4dcc6a68-7a7e-4a85-8566-794266e676cd tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 888.649036] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-4dcc6a68-7a7e-4a85-8566-794266e676cd tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Deleted contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 888.649036] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-4dcc6a68-7a7e-4a85-8566-794266e676cd tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 888.649248] env[61868]: INFO nova.compute.manager [None req-4dcc6a68-7a7e-4a85-8566-794266e676cd tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Took 1.15 seconds to destroy the instance on the hypervisor. [ 888.649358] env[61868]: DEBUG oslo.service.loopingcall [None req-4dcc6a68-7a7e-4a85-8566-794266e676cd tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 888.649535] env[61868]: DEBUG nova.compute.manager [-] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 888.649630] env[61868]: DEBUG nova.network.neutron [-] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 888.659642] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1f690bc9-2e92-4774-a822-25d8505bdac3 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Acquiring lock "083140bf-233d-49ac-8247-cd93206f9200" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.687433] env[61868]: DEBUG nova.network.neutron [-] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.734583] env[61868]: DEBUG nova.network.neutron [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 888.763420] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-32125dfb-30bb-491f-bc72-5a350db11c9a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.774883] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7870742-83d3-4a06-9903-38afd2aa4022 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.802340] env[61868]: DEBUG nova.compute.manager [req-8fbc765a-a49d-43d6-81f7-e10d67e9666d req-5d1c4e26-33a3-43f5-8750-a8866f52edda service nova] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Detach interface failed, port_id=ae9b8f9e-17ec-4c1f-8052-8467a711410c, reason: Instance b9b5be37-6b30-4229-9c8d-3ee9d30db119 could not be found. {{(pid=61868) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 888.891284] env[61868]: DEBUG nova.compute.manager [req-507eede8-e9d4-4074-b238-6939fdb58468 req-f3d5ec5c-3736-4851-aad0-260ebe8ddc73 service nova] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Received event network-vif-plugged-4186aca7-63b5-4b24-ab7c-de8f87064242 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 888.893470] env[61868]: DEBUG oslo_concurrency.lockutils [req-507eede8-e9d4-4074-b238-6939fdb58468 req-f3d5ec5c-3736-4851-aad0-260ebe8ddc73 service nova] Acquiring lock "306f2dc4-933c-4948-8330-eda93931c4e0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.893470] env[61868]: DEBUG oslo_concurrency.lockutils [req-507eede8-e9d4-4074-b238-6939fdb58468 req-f3d5ec5c-3736-4851-aad0-260ebe8ddc73 service nova] Lock "306f2dc4-933c-4948-8330-eda93931c4e0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.893470] env[61868]: DEBUG oslo_concurrency.lockutils [req-507eede8-e9d4-4074-b238-6939fdb58468 req-f3d5ec5c-3736-4851-aad0-260ebe8ddc73 service nova] Lock "306f2dc4-933c-4948-8330-eda93931c4e0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.893470] env[61868]: DEBUG nova.compute.manager [req-507eede8-e9d4-4074-b238-6939fdb58468 req-f3d5ec5c-3736-4851-aad0-260ebe8ddc73 service nova] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] No waiting events found dispatching network-vif-plugged-4186aca7-63b5-4b24-ab7c-de8f87064242 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 888.893470] env[61868]: WARNING nova.compute.manager [req-507eede8-e9d4-4074-b238-6939fdb58468 req-f3d5ec5c-3736-4851-aad0-260ebe8ddc73 service nova] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Received unexpected event network-vif-plugged-4186aca7-63b5-4b24-ab7c-de8f87064242 for instance with vm_state building and task_state spawning. [ 888.893632] env[61868]: DEBUG nova.compute.manager [req-507eede8-e9d4-4074-b238-6939fdb58468 req-f3d5ec5c-3736-4851-aad0-260ebe8ddc73 service nova] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Received event network-changed-4186aca7-63b5-4b24-ab7c-de8f87064242 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 888.893632] env[61868]: DEBUG nova.compute.manager [req-507eede8-e9d4-4074-b238-6939fdb58468 req-f3d5ec5c-3736-4851-aad0-260ebe8ddc73 service nova] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Refreshing instance network info cache due to event network-changed-4186aca7-63b5-4b24-ab7c-de8f87064242. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 888.893632] env[61868]: DEBUG oslo_concurrency.lockutils [req-507eede8-e9d4-4074-b238-6939fdb58468 req-f3d5ec5c-3736-4851-aad0-260ebe8ddc73 service nova] Acquiring lock "refresh_cache-306f2dc4-933c-4948-8330-eda93931c4e0" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.894157] env[61868]: DEBUG oslo_concurrency.lockutils [None req-fc5198bf-c3fb-4340-a65e-568a649e95ca tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.896397] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.860s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.899037] env[61868]: INFO nova.compute.claims [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 888.985281] env[61868]: DEBUG nova.network.neutron [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Updating instance_info_cache with network_info: [{"id": "4186aca7-63b5-4b24-ab7c-de8f87064242", "address": "fa:16:3e:da:51:1e", "network": {"id": "f2e3f911-d826-43ad-911d-9f6985db7ee5", "bridge": "br-int", "label": "tempest-ImagesTestJSON-485234144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5658848e1b0e42929c04a4a8de40a291", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4186aca7-63", "ovs_interfaceid": "4186aca7-63b5-4b24-ab7c-de8f87064242", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.033951] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-805d5425-8b50-412b-81b4-c3b60c28ab13 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 889.034298] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-805d5425-8b50-412b-81b4-c3b60c28ab13 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 889.034516] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-805d5425-8b50-412b-81b4-c3b60c28ab13 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Deleting the datastore file [datastore1] b7055e5a-37d2-42d5-bab0-1e70faf52b58 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 889.034825] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e8501306-b761-4c0f-8fbe-52fce1f00ca8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.043778] env[61868]: DEBUG oslo_vmware.api [None req-805d5425-8b50-412b-81b4-c3b60c28ab13 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 889.043778] env[61868]: value = "task-1315579" [ 889.043778] env[61868]: _type = "Task" [ 889.043778] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.055269] env[61868]: DEBUG oslo_vmware.api [None req-805d5425-8b50-412b-81b4-c3b60c28ab13 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315579, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.142026] env[61868]: DEBUG oslo_vmware.api [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': task-1315577, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.189129] env[61868]: INFO nova.compute.manager [-] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Took 1.52 seconds to deallocate network for instance. [ 889.426730] env[61868]: DEBUG nova.network.neutron [-] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.490018] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Releasing lock "refresh_cache-306f2dc4-933c-4948-8330-eda93931c4e0" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.490018] env[61868]: DEBUG nova.compute.manager [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Instance network_info: |[{"id": "4186aca7-63b5-4b24-ab7c-de8f87064242", "address": "fa:16:3e:da:51:1e", "network": {"id": "f2e3f911-d826-43ad-911d-9f6985db7ee5", "bridge": "br-int", "label": "tempest-ImagesTestJSON-485234144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5658848e1b0e42929c04a4a8de40a291", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4186aca7-63", "ovs_interfaceid": "4186aca7-63b5-4b24-ab7c-de8f87064242", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 889.490309] env[61868]: DEBUG oslo_concurrency.lockutils [req-507eede8-e9d4-4074-b238-6939fdb58468 req-f3d5ec5c-3736-4851-aad0-260ebe8ddc73 service nova] Acquired lock "refresh_cache-306f2dc4-933c-4948-8330-eda93931c4e0" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.490309] env[61868]: DEBUG nova.network.neutron [req-507eede8-e9d4-4074-b238-6939fdb58468 req-f3d5ec5c-3736-4851-aad0-260ebe8ddc73 service nova] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Refreshing network info cache for port 4186aca7-63b5-4b24-ab7c-de8f87064242 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 889.490657] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:da:51:1e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f1b507ed-cd2d-4c09-9d96-c47bde6a7774', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4186aca7-63b5-4b24-ab7c-de8f87064242', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 889.498525] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Creating folder: Project (5658848e1b0e42929c04a4a8de40a291). Parent ref: group-v281478. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 889.499404] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-192df4ea-a9e8-40fc-b7eb-43973759c143 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.516036] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Created folder: Project (5658848e1b0e42929c04a4a8de40a291) in parent group-v281478. [ 889.516334] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Creating folder: Instances. Parent ref: group-v281557. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 889.516608] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d822fc70-4803-4651-8970-937f9473a641 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.531308] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Created folder: Instances in parent group-v281557. [ 889.531308] env[61868]: DEBUG oslo.service.loopingcall [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 889.531308] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 889.531308] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c585553e-11bd-495b-beb8-d4c9a5574eaf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.554439] env[61868]: DEBUG oslo_vmware.api [None req-805d5425-8b50-412b-81b4-c3b60c28ab13 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315579, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141328} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.555707] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-805d5425-8b50-412b-81b4-c3b60c28ab13 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 889.555934] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-805d5425-8b50-412b-81b4-c3b60c28ab13 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 889.556193] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-805d5425-8b50-412b-81b4-c3b60c28ab13 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 889.556790] env[61868]: INFO nova.compute.manager [None req-805d5425-8b50-412b-81b4-c3b60c28ab13 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Took 2.22 seconds to destroy the instance on the hypervisor. [ 889.556790] env[61868]: DEBUG oslo.service.loopingcall [None req-805d5425-8b50-412b-81b4-c3b60c28ab13 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 889.556902] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 889.556902] env[61868]: value = "task-1315582" [ 889.556902] env[61868]: _type = "Task" [ 889.556902] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.557110] env[61868]: DEBUG nova.compute.manager [-] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 889.557384] env[61868]: DEBUG nova.network.neutron [-] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 889.568154] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315582, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.642174] env[61868]: DEBUG oslo_vmware.api [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': task-1315577, 'name': ReconfigVM_Task, 'duration_secs': 1.238169} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.642478] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Reconfigured VM instance instance-0000004a to attach disk [datastore1] f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3/f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 889.643469] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-55c7646e-9b23-4955-a15b-1be1eeabc4fd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.651128] env[61868]: DEBUG oslo_vmware.api [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Waiting for the task: (returnval){ [ 889.651128] env[61868]: value = "task-1315583" [ 889.651128] env[61868]: _type = "Task" [ 889.651128] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.661844] env[61868]: DEBUG oslo_vmware.api [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': task-1315583, 'name': Rename_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.677470] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Acquiring lock "4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.677617] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Lock "4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.698463] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0f7f9081-032d-4547-939e-b812c9bdc0c7 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.930742] env[61868]: INFO nova.compute.manager [-] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Took 1.28 seconds to deallocate network for instance. [ 890.077318] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315582, 'name': CreateVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.165145] env[61868]: DEBUG oslo_vmware.api [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': task-1315583, 'name': Rename_Task, 'duration_secs': 0.15357} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.165493] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 890.165740] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1f0550dc-3e9b-43e3-b090-cf7a4e23b5be {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.174200] env[61868]: DEBUG oslo_vmware.api [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Waiting for the task: (returnval){ [ 890.174200] env[61868]: value = "task-1315584" [ 890.174200] env[61868]: _type = "Task" [ 890.174200] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.182572] env[61868]: DEBUG nova.compute.manager [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 890.191542] env[61868]: DEBUG oslo_vmware.api [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': task-1315584, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.278516] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9aad792-23c5-4de5-adc9-05f30a663bc6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.290740] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e654af8a-7bfe-4170-93f3-47e7b6a8ee55 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.328304] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5fba8c4-7df3-459e-9823-35a163343e24 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.338467] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a695531-1fb9-4da3-914a-eb804669e784 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.343566] env[61868]: DEBUG nova.compute.manager [req-92a3cba9-bfba-4f0b-9841-6f36b0a2b4da req-78d6f506-8e44-4755-bafe-60306c2e3467 service nova] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Received event network-vif-deleted-981ca2e3-c2d7-4d28-bda6-68ddd6d57bc5 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 890.354463] env[61868]: DEBUG nova.compute.provider_tree [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 890.424536] env[61868]: DEBUG nova.network.neutron [-] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.437972] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4dcc6a68-7a7e-4a85-8566-794266e676cd tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.508753] env[61868]: DEBUG nova.network.neutron [req-507eede8-e9d4-4074-b238-6939fdb58468 req-f3d5ec5c-3736-4851-aad0-260ebe8ddc73 service nova] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Updated VIF entry in instance network info cache for port 4186aca7-63b5-4b24-ab7c-de8f87064242. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 890.509135] env[61868]: DEBUG nova.network.neutron [req-507eede8-e9d4-4074-b238-6939fdb58468 req-f3d5ec5c-3736-4851-aad0-260ebe8ddc73 service nova] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Updating instance_info_cache with network_info: [{"id": "4186aca7-63b5-4b24-ab7c-de8f87064242", "address": "fa:16:3e:da:51:1e", "network": {"id": "f2e3f911-d826-43ad-911d-9f6985db7ee5", "bridge": "br-int", "label": "tempest-ImagesTestJSON-485234144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5658848e1b0e42929c04a4a8de40a291", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4186aca7-63", "ovs_interfaceid": "4186aca7-63b5-4b24-ab7c-de8f87064242", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.570492] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315582, 'name': CreateVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.691543] env[61868]: DEBUG oslo_vmware.api [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': task-1315584, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.707369] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.862133] env[61868]: DEBUG nova.scheduler.client.report [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 890.915202] env[61868]: DEBUG nova.compute.manager [req-c4c9faae-f44e-4fe9-a2e7-23ab3c563ff8 req-5e9b55af-0cc5-4f99-809d-83fb76b0f9c5 service nova] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Received event network-vif-deleted-8ca7ca99-eddc-41fa-bfbc-1fafcaf733cc {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 890.927058] env[61868]: INFO nova.compute.manager [-] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Took 1.37 seconds to deallocate network for instance. [ 891.012195] env[61868]: DEBUG oslo_concurrency.lockutils [req-507eede8-e9d4-4074-b238-6939fdb58468 req-f3d5ec5c-3736-4851-aad0-260ebe8ddc73 service nova] Releasing lock "refresh_cache-306f2dc4-933c-4948-8330-eda93931c4e0" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.070122] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315582, 'name': CreateVM_Task, 'duration_secs': 1.345297} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.070280] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 891.070989] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.071183] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.071504] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 891.071797] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4ec2834-dc07-4f3c-a456-3dbffe13bb54 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.076296] env[61868]: DEBUG oslo_vmware.api [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 891.076296] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]525ab39f-93ce-11d9-08ba-ff19002d79a1" [ 891.076296] env[61868]: _type = "Task" [ 891.076296] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.083540] env[61868]: DEBUG oslo_vmware.api [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]525ab39f-93ce-11d9-08ba-ff19002d79a1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.187494] env[61868]: DEBUG oslo_vmware.api [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': task-1315584, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.365707] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.470s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.366276] env[61868]: DEBUG nova.compute.manager [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 891.368996] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a1b309cb-b151-464d-8aab-50c6857b088e tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.750s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.369246] env[61868]: DEBUG nova.objects.instance [None req-a1b309cb-b151-464d-8aab-50c6857b088e tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Lazy-loading 'resources' on Instance uuid c6f99cbc-3e81-44b4-b3d1-d8cd821829f2 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 891.434034] env[61868]: DEBUG oslo_concurrency.lockutils [None req-805d5425-8b50-412b-81b4-c3b60c28ab13 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.587185] env[61868]: DEBUG oslo_vmware.api [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]525ab39f-93ce-11d9-08ba-ff19002d79a1, 'name': SearchDatastore_Task, 'duration_secs': 0.008859} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.587501] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.587735] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 891.587969] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.588137] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.588322] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 891.588580] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-682e9dfd-1676-4176-b93d-ad1a9e08b46e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.597390] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 891.597551] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 891.598267] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7531db5d-157b-470b-b821-62d20493fbea {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.604107] env[61868]: DEBUG oslo_vmware.api [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 891.604107] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5224c0d5-b84f-e8c2-d04b-a16b87afa922" [ 891.604107] env[61868]: _type = "Task" [ 891.604107] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.611432] env[61868]: DEBUG oslo_vmware.api [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5224c0d5-b84f-e8c2-d04b-a16b87afa922, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.687970] env[61868]: DEBUG oslo_vmware.api [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': task-1315584, 'name': PowerOnVM_Task, 'duration_secs': 1.04137} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.688264] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 891.688469] env[61868]: INFO nova.compute.manager [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Took 9.66 seconds to spawn the instance on the hypervisor. [ 891.688645] env[61868]: DEBUG nova.compute.manager [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 891.689404] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3971fdab-a63e-4405-9402-bc5a1f54e917 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.872684] env[61868]: DEBUG nova.compute.utils [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 891.876991] env[61868]: DEBUG nova.compute.manager [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 891.877172] env[61868]: DEBUG nova.network.neutron [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 891.958377] env[61868]: DEBUG nova.policy [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3c44b4da602b4022b643fc49d72f2d22', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2b9b800086764b41a8817d1d8e463a2f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 892.116436] env[61868]: DEBUG oslo_vmware.api [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5224c0d5-b84f-e8c2-d04b-a16b87afa922, 'name': SearchDatastore_Task, 'duration_secs': 0.00876} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.117318] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7648448d-a51d-4b15-be76-6feb93d56f86 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.123674] env[61868]: DEBUG oslo_vmware.api [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 892.123674] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5277bac0-44d1-0358-83ac-aefd86a7cf12" [ 892.123674] env[61868]: _type = "Task" [ 892.123674] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.132931] env[61868]: DEBUG oslo_vmware.api [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5277bac0-44d1-0358-83ac-aefd86a7cf12, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.179866] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9eacfae-8641-4e6f-b081-5185e4dbd1b3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.188912] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c1a011c-fef2-442e-a9a2-fa3bc2c01df4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.225587] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c8e7688-2d8f-4762-9c17-f7c518e35cbd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.230659] env[61868]: INFO nova.compute.manager [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Took 32.70 seconds to build instance. [ 892.236497] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a07875d4-d749-4ac5-b378-15e7b9f94b09 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.252814] env[61868]: DEBUG nova.compute.provider_tree [None req-a1b309cb-b151-464d-8aab-50c6857b088e tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 892.264653] env[61868]: DEBUG nova.network.neutron [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Successfully created port: e76c5ee1-88d7-4366-a731-7b793639fcef {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 892.378077] env[61868]: DEBUG nova.compute.manager [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 892.636234] env[61868]: DEBUG oslo_vmware.api [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5277bac0-44d1-0358-83ac-aefd86a7cf12, 'name': SearchDatastore_Task, 'duration_secs': 0.011897} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.636510] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.636767] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 306f2dc4-933c-4948-8330-eda93931c4e0/306f2dc4-933c-4948-8330-eda93931c4e0.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 892.637111] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-35187f04-0860-42c5-9b3c-cf4db08cde26 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.644497] env[61868]: DEBUG oslo_vmware.api [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 892.644497] env[61868]: value = "task-1315585" [ 892.644497] env[61868]: _type = "Task" [ 892.644497] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.655011] env[61868]: DEBUG oslo_vmware.api [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315585, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.733050] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2b801008-2112-433b-b58a-f7dbe354557f tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Lock "f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 62.722s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.755647] env[61868]: DEBUG nova.scheduler.client.report [None req-a1b309cb-b151-464d-8aab-50c6857b088e tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 893.155573] env[61868]: DEBUG oslo_vmware.api [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315585, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.472454} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.155881] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 306f2dc4-933c-4948-8330-eda93931c4e0/306f2dc4-933c-4948-8330-eda93931c4e0.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 893.156115] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 893.156378] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-da6e4032-7cd1-4244-a072-53aa4b978acb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.163359] env[61868]: DEBUG oslo_vmware.api [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 893.163359] env[61868]: value = "task-1315586" [ 893.163359] env[61868]: _type = "Task" [ 893.163359] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.171951] env[61868]: DEBUG oslo_vmware.api [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315586, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.260666] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a1b309cb-b151-464d-8aab-50c6857b088e tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.892s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.263063] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.152s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.264547] env[61868]: INFO nova.compute.claims [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 893.286693] env[61868]: INFO nova.scheduler.client.report [None req-a1b309cb-b151-464d-8aab-50c6857b088e tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Deleted allocations for instance c6f99cbc-3e81-44b4-b3d1-d8cd821829f2 [ 893.294103] env[61868]: INFO nova.compute.manager [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Rescuing [ 893.295086] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Acquiring lock "refresh_cache-f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.295221] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Acquired lock "refresh_cache-f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.295804] env[61868]: DEBUG nova.network.neutron [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 893.389023] env[61868]: DEBUG nova.compute.manager [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 893.415425] env[61868]: DEBUG nova.virt.hardware [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 893.415696] env[61868]: DEBUG nova.virt.hardware [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 893.415852] env[61868]: DEBUG nova.virt.hardware [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 893.416051] env[61868]: DEBUG nova.virt.hardware [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 893.416211] env[61868]: DEBUG nova.virt.hardware [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 893.416361] env[61868]: DEBUG nova.virt.hardware [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 893.416576] env[61868]: DEBUG nova.virt.hardware [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 893.416739] env[61868]: DEBUG nova.virt.hardware [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 893.416934] env[61868]: DEBUG nova.virt.hardware [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 893.417138] env[61868]: DEBUG nova.virt.hardware [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 893.417324] env[61868]: DEBUG nova.virt.hardware [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 893.418348] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2615a945-e735-4bf1-a643-9787538d425c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.427676] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55950618-46ed-43ad-84b4-334b9ffcb792 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.675163] env[61868]: DEBUG oslo_vmware.api [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315586, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064754} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.675538] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 893.676677] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc112cc1-c0b6-4ad3-b003-65a4df568b2e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.702698] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] 306f2dc4-933c-4948-8330-eda93931c4e0/306f2dc4-933c-4948-8330-eda93931c4e0.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 893.703103] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5d0f99ce-4996-4bc2-aed4-b88437ccb2c8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.727879] env[61868]: DEBUG oslo_vmware.api [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 893.727879] env[61868]: value = "task-1315587" [ 893.727879] env[61868]: _type = "Task" [ 893.727879] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.739267] env[61868]: DEBUG oslo_vmware.api [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315587, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.794601] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a1b309cb-b151-464d-8aab-50c6857b088e tempest-ServerMetadataTestJSON-1107515565 tempest-ServerMetadataTestJSON-1107515565-project-member] Lock "c6f99cbc-3e81-44b4-b3d1-d8cd821829f2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.229s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.043136] env[61868]: DEBUG nova.network.neutron [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Updating instance_info_cache with network_info: [{"id": "29e8f7bd-db77-4071-b4c6-79e8338c2e82", "address": "fa:16:3e:be:13:44", "network": {"id": "4befda35-5c88-4ac7-b8ef-128ddae97a7d", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-113207814-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "78a206c3daa14cc58e7fe8546f565ae0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd72ef32-a57c-43b0-93df-e8a030987d44", "external-id": "nsx-vlan-transportzone-340", "segmentation_id": 340, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap29e8f7bd-db", "ovs_interfaceid": "29e8f7bd-db77-4071-b4c6-79e8338c2e82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.105887] env[61868]: DEBUG nova.compute.manager [req-0440a269-7a68-498c-b078-58ef5298d751 req-8ac225b4-bbba-4c0d-83d3-987fe645f905 service nova] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Received event network-vif-plugged-e76c5ee1-88d7-4366-a731-7b793639fcef {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 894.106134] env[61868]: DEBUG oslo_concurrency.lockutils [req-0440a269-7a68-498c-b078-58ef5298d751 req-8ac225b4-bbba-4c0d-83d3-987fe645f905 service nova] Acquiring lock "e2832c0f-dc62-4628-87d4-e2a01819b771-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.106348] env[61868]: DEBUG oslo_concurrency.lockutils [req-0440a269-7a68-498c-b078-58ef5298d751 req-8ac225b4-bbba-4c0d-83d3-987fe645f905 service nova] Lock "e2832c0f-dc62-4628-87d4-e2a01819b771-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.106593] env[61868]: DEBUG oslo_concurrency.lockutils [req-0440a269-7a68-498c-b078-58ef5298d751 req-8ac225b4-bbba-4c0d-83d3-987fe645f905 service nova] Lock "e2832c0f-dc62-4628-87d4-e2a01819b771-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.106698] env[61868]: DEBUG nova.compute.manager [req-0440a269-7a68-498c-b078-58ef5298d751 req-8ac225b4-bbba-4c0d-83d3-987fe645f905 service nova] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] No waiting events found dispatching network-vif-plugged-e76c5ee1-88d7-4366-a731-7b793639fcef {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 894.106928] env[61868]: WARNING nova.compute.manager [req-0440a269-7a68-498c-b078-58ef5298d751 req-8ac225b4-bbba-4c0d-83d3-987fe645f905 service nova] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Received unexpected event network-vif-plugged-e76c5ee1-88d7-4366-a731-7b793639fcef for instance with vm_state building and task_state spawning. [ 894.240944] env[61868]: DEBUG oslo_vmware.api [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315587, 'name': ReconfigVM_Task, 'duration_secs': 0.28765} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.240944] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Reconfigured VM instance instance-0000004b to attach disk [datastore1] 306f2dc4-933c-4948-8330-eda93931c4e0/306f2dc4-933c-4948-8330-eda93931c4e0.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 894.240944] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-da98e748-a309-4d3e-84e8-d0d48ba5484c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.249022] env[61868]: DEBUG oslo_vmware.api [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 894.249022] env[61868]: value = "task-1315588" [ 894.249022] env[61868]: _type = "Task" [ 894.249022] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.255797] env[61868]: DEBUG oslo_vmware.api [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315588, 'name': Rename_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.290072] env[61868]: DEBUG nova.network.neutron [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Successfully updated port: e76c5ee1-88d7-4366-a731-7b793639fcef {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 894.546010] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Releasing lock "refresh_cache-f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.638512] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f44397f-157f-4350-bcdd-58c78abc7dfc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.647546] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-947ff72c-8bf1-4e8f-a396-afcef09feca5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.684448] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d87e13b9-ef66-4470-8495-18207df78f75 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.693321] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb97e616-7ef9-46e0-a9d4-ce2fc283a269 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.714097] env[61868]: DEBUG nova.compute.provider_tree [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 894.759766] env[61868]: DEBUG oslo_vmware.api [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315588, 'name': Rename_Task, 'duration_secs': 0.148595} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.759766] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 894.759766] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-409393e6-0bab-40be-a70c-0f49b7b0280d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.772577] env[61868]: DEBUG oslo_vmware.api [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 894.772577] env[61868]: value = "task-1315589" [ 894.772577] env[61868]: _type = "Task" [ 894.772577] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.785302] env[61868]: DEBUG oslo_vmware.api [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315589, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.791629] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquiring lock "refresh_cache-e2832c0f-dc62-4628-87d4-e2a01819b771" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.793199] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquired lock "refresh_cache-e2832c0f-dc62-4628-87d4-e2a01819b771" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.793199] env[61868]: DEBUG nova.network.neutron [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 895.085249] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 895.085534] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f402d023-345a-44b4-bbf3-8c65fbf0bcf7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.094490] env[61868]: DEBUG oslo_vmware.api [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Waiting for the task: (returnval){ [ 895.094490] env[61868]: value = "task-1315590" [ 895.094490] env[61868]: _type = "Task" [ 895.094490] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.104764] env[61868]: DEBUG oslo_vmware.api [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': task-1315590, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.218865] env[61868]: DEBUG nova.scheduler.client.report [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 895.285523] env[61868]: DEBUG oslo_vmware.api [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315589, 'name': PowerOnVM_Task, 'duration_secs': 0.467901} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.285960] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 895.286118] env[61868]: INFO nova.compute.manager [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Took 8.86 seconds to spawn the instance on the hypervisor. [ 895.286348] env[61868]: DEBUG nova.compute.manager [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 895.287224] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7d0d3be-0a24-4c2d-ab3d-9b41b177a2c2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.400702] env[61868]: DEBUG nova.network.neutron [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 895.605571] env[61868]: DEBUG oslo_vmware.api [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': task-1315590, 'name': PowerOffVM_Task, 'duration_secs': 0.335876} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.607979] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 895.608904] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1740fd3-3736-44e0-bd2a-f4df3311f87a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.629091] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dab48460-4b9e-48dc-80c3-5cb161fd6e87 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.669525] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 895.669890] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-67a3207b-dcff-44f4-92de-558b45698d25 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.672384] env[61868]: DEBUG nova.network.neutron [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Updating instance_info_cache with network_info: [{"id": "e76c5ee1-88d7-4366-a731-7b793639fcef", "address": "fa:16:3e:5b:e3:3f", "network": {"id": "e7457f1a-171a-4c82-a260-1d1a7a95861e", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1032778554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b9b800086764b41a8817d1d8e463a2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape76c5ee1-88", "ovs_interfaceid": "e76c5ee1-88d7-4366-a731-7b793639fcef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.680275] env[61868]: DEBUG oslo_vmware.api [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Waiting for the task: (returnval){ [ 895.680275] env[61868]: value = "task-1315591" [ 895.680275] env[61868]: _type = "Task" [ 895.680275] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.692221] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] VM already powered off {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 895.692468] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 895.692712] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.692862] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.693089] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 895.694066] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ec71ba86-68f4-4fea-a9b6-cf866efc8200 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.702372] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 895.705021] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 895.705021] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-44a29311-49ec-4848-9f8e-fa2756acff3d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.710122] env[61868]: DEBUG oslo_vmware.api [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Waiting for the task: (returnval){ [ 895.710122] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5220b2d5-4d68-2f8a-a899-8c1ae707e5f3" [ 895.710122] env[61868]: _type = "Task" [ 895.710122] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.718860] env[61868]: DEBUG oslo_vmware.api [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5220b2d5-4d68-2f8a-a899-8c1ae707e5f3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.724098] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.461s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.725027] env[61868]: DEBUG nova.compute.manager [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 895.727085] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.338s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.728589] env[61868]: INFO nova.compute.claims [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 895.813537] env[61868]: INFO nova.compute.manager [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Took 32.79 seconds to build instance. [ 896.042403] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4913bdaa-6bad-444f-8690-03eeb16ef4b9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Acquiring lock "409cad1a-946d-4c58-aa57-1c0bf97fe63f" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.042513] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4913bdaa-6bad-444f-8690-03eeb16ef4b9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Lock "409cad1a-946d-4c58-aa57-1c0bf97fe63f" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.175056] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Releasing lock "refresh_cache-e2832c0f-dc62-4628-87d4-e2a01819b771" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.175313] env[61868]: DEBUG nova.compute.manager [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Instance network_info: |[{"id": "e76c5ee1-88d7-4366-a731-7b793639fcef", "address": "fa:16:3e:5b:e3:3f", "network": {"id": "e7457f1a-171a-4c82-a260-1d1a7a95861e", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1032778554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b9b800086764b41a8817d1d8e463a2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape76c5ee1-88", "ovs_interfaceid": "e76c5ee1-88d7-4366-a731-7b793639fcef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 896.176368] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5b:e3:3f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd19577c9-1b2e-490b-8031-2f278dd3f570', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e76c5ee1-88d7-4366-a731-7b793639fcef', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 896.184928] env[61868]: DEBUG oslo.service.loopingcall [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 896.186573] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 896.187733] env[61868]: DEBUG nova.compute.manager [req-7dd96a93-43e6-4898-9626-c30be77e98db req-ee80426a-fa76-42e9-be38-fe4014d2eec3 service nova] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Received event network-changed-e76c5ee1-88d7-4366-a731-7b793639fcef {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 896.187948] env[61868]: DEBUG nova.compute.manager [req-7dd96a93-43e6-4898-9626-c30be77e98db req-ee80426a-fa76-42e9-be38-fe4014d2eec3 service nova] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Refreshing instance network info cache due to event network-changed-e76c5ee1-88d7-4366-a731-7b793639fcef. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 896.188185] env[61868]: DEBUG oslo_concurrency.lockutils [req-7dd96a93-43e6-4898-9626-c30be77e98db req-ee80426a-fa76-42e9-be38-fe4014d2eec3 service nova] Acquiring lock "refresh_cache-e2832c0f-dc62-4628-87d4-e2a01819b771" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.188329] env[61868]: DEBUG oslo_concurrency.lockutils [req-7dd96a93-43e6-4898-9626-c30be77e98db req-ee80426a-fa76-42e9-be38-fe4014d2eec3 service nova] Acquired lock "refresh_cache-e2832c0f-dc62-4628-87d4-e2a01819b771" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.188496] env[61868]: DEBUG nova.network.neutron [req-7dd96a93-43e6-4898-9626-c30be77e98db req-ee80426a-fa76-42e9-be38-fe4014d2eec3 service nova] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Refreshing network info cache for port e76c5ee1-88d7-4366-a731-7b793639fcef {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 896.190712] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-da1d442a-218a-43cc-86f2-391c58e4c128 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.215808] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 896.215808] env[61868]: value = "task-1315592" [ 896.215808] env[61868]: _type = "Task" [ 896.215808] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.223315] env[61868]: DEBUG oslo_vmware.api [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5220b2d5-4d68-2f8a-a899-8c1ae707e5f3, 'name': SearchDatastore_Task, 'duration_secs': 0.008854} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.224695] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e37b7a0-1a0f-4816-a67a-56f56b9b740d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.230443] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315592, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.233505] env[61868]: DEBUG nova.compute.utils [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 896.240505] env[61868]: DEBUG nova.compute.manager [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 896.240712] env[61868]: DEBUG nova.network.neutron [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 896.242734] env[61868]: DEBUG oslo_vmware.api [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Waiting for the task: (returnval){ [ 896.242734] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]520a9d40-9666-6a78-1203-d34898f50585" [ 896.242734] env[61868]: _type = "Task" [ 896.242734] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.261192] env[61868]: DEBUG oslo_vmware.api [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]520a9d40-9666-6a78-1203-d34898f50585, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.299092] env[61868]: DEBUG nova.policy [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1425850b70c940988169260ea59c21df', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7541757d43c74b93acf728aa2fb0f425', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 896.317081] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a24a546d-9972-4ea9-8647-c9bf2c82e657 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "306f2dc4-933c-4948-8330-eda93931c4e0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 58.908s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.389930] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f70d8e25-e795-41bd-ae53-9acba3f1fd71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "306f2dc4-933c-4948-8330-eda93931c4e0" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.390229] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f70d8e25-e795-41bd-ae53-9acba3f1fd71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "306f2dc4-933c-4948-8330-eda93931c4e0" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.390420] env[61868]: DEBUG nova.compute.manager [None req-f70d8e25-e795-41bd-ae53-9acba3f1fd71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 896.391390] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-905b2d4d-599e-46dc-8daf-9ab54739f9a9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.398670] env[61868]: DEBUG nova.compute.manager [None req-f70d8e25-e795-41bd-ae53-9acba3f1fd71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61868) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 896.400389] env[61868]: DEBUG nova.objects.instance [None req-f70d8e25-e795-41bd-ae53-9acba3f1fd71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lazy-loading 'flavor' on Instance uuid 306f2dc4-933c-4948-8330-eda93931c4e0 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 896.545631] env[61868]: DEBUG nova.compute.utils [None req-4913bdaa-6bad-444f-8690-03eeb16ef4b9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 896.642842] env[61868]: DEBUG nova.network.neutron [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Successfully created port: c27d3e06-cf3c-4ef8-ba0b-e6f2ca824646 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 896.727975] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315592, 'name': CreateVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.740952] env[61868]: DEBUG nova.compute.manager [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 896.758355] env[61868]: DEBUG oslo_vmware.api [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]520a9d40-9666-6a78-1203-d34898f50585, 'name': SearchDatastore_Task, 'duration_secs': 0.025694} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.758904] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.759187] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d-rescue.vmdk. {{(pid=61868) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 896.759458] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b84e2297-3bd4-4dff-91a3-2b521b9f3e7a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.768321] env[61868]: DEBUG oslo_vmware.api [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Waiting for the task: (returnval){ [ 896.768321] env[61868]: value = "task-1315593" [ 896.768321] env[61868]: _type = "Task" [ 896.768321] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.786952] env[61868]: DEBUG oslo_vmware.api [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': task-1315593, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.905466] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-f70d8e25-e795-41bd-ae53-9acba3f1fd71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 896.905748] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bfde5ac7-5bc8-453d-837a-6c4f6fcbb73b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.914076] env[61868]: DEBUG oslo_vmware.api [None req-f70d8e25-e795-41bd-ae53-9acba3f1fd71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 896.914076] env[61868]: value = "task-1315594" [ 896.914076] env[61868]: _type = "Task" [ 896.914076] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.923049] env[61868]: DEBUG oslo_vmware.api [None req-f70d8e25-e795-41bd-ae53-9acba3f1fd71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315594, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.938145] env[61868]: DEBUG nova.network.neutron [req-7dd96a93-43e6-4898-9626-c30be77e98db req-ee80426a-fa76-42e9-be38-fe4014d2eec3 service nova] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Updated VIF entry in instance network info cache for port e76c5ee1-88d7-4366-a731-7b793639fcef. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 896.938553] env[61868]: DEBUG nova.network.neutron [req-7dd96a93-43e6-4898-9626-c30be77e98db req-ee80426a-fa76-42e9-be38-fe4014d2eec3 service nova] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Updating instance_info_cache with network_info: [{"id": "e76c5ee1-88d7-4366-a731-7b793639fcef", "address": "fa:16:3e:5b:e3:3f", "network": {"id": "e7457f1a-171a-4c82-a260-1d1a7a95861e", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1032778554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b9b800086764b41a8817d1d8e463a2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape76c5ee1-88", "ovs_interfaceid": "e76c5ee1-88d7-4366-a731-7b793639fcef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.048111] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4913bdaa-6bad-444f-8690-03eeb16ef4b9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Lock "409cad1a-946d-4c58-aa57-1c0bf97fe63f" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.005s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.051410] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80bc14c6-394f-4636-aa07-89dca0da2c93 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.060128] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47595c18-dada-4294-8a02-732e42876046 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.091744] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89fd8cd8-0774-4537-beca-8a1caacd80c4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.099891] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4d4cf27-9cac-442f-a00c-7e2945da6511 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.113536] env[61868]: DEBUG nova.compute.provider_tree [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 897.227795] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315592, 'name': CreateVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.278309] env[61868]: DEBUG oslo_vmware.api [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': task-1315593, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.424833] env[61868]: DEBUG oslo_vmware.api [None req-f70d8e25-e795-41bd-ae53-9acba3f1fd71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315594, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.443584] env[61868]: DEBUG oslo_concurrency.lockutils [req-7dd96a93-43e6-4898-9626-c30be77e98db req-ee80426a-fa76-42e9-be38-fe4014d2eec3 service nova] Releasing lock "refresh_cache-e2832c0f-dc62-4628-87d4-e2a01819b771" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.616905] env[61868]: DEBUG nova.scheduler.client.report [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 897.730993] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315592, 'name': CreateVM_Task, 'duration_secs': 1.490528} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.731197] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 897.731987] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.732200] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.732520] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 897.732777] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a5509d52-00d0-4fa0-8370-ee5e5f2f2102 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.738018] env[61868]: DEBUG oslo_vmware.api [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the task: (returnval){ [ 897.738018] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]521724df-44d6-4479-4f73-022c4ddb9b91" [ 897.738018] env[61868]: _type = "Task" [ 897.738018] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.746685] env[61868]: DEBUG oslo_vmware.api [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]521724df-44d6-4479-4f73-022c4ddb9b91, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.751336] env[61868]: DEBUG nova.compute.manager [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 897.781344] env[61868]: DEBUG oslo_vmware.api [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': task-1315593, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.784010] env[61868]: DEBUG nova.virt.hardware [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 897.784320] env[61868]: DEBUG nova.virt.hardware [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 897.784623] env[61868]: DEBUG nova.virt.hardware [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 897.784792] env[61868]: DEBUG nova.virt.hardware [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 897.784964] env[61868]: DEBUG nova.virt.hardware [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 897.785188] env[61868]: DEBUG nova.virt.hardware [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 897.786321] env[61868]: DEBUG nova.virt.hardware [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 897.786321] env[61868]: DEBUG nova.virt.hardware [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 897.786321] env[61868]: DEBUG nova.virt.hardware [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 897.786321] env[61868]: DEBUG nova.virt.hardware [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 897.786600] env[61868]: DEBUG nova.virt.hardware [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 897.787530] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd538174-b715-4ac5-8cb4-00c78e33360e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.802780] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58ea66c3-3159-4caa-89a9-76babafa457b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.928109] env[61868]: DEBUG oslo_vmware.api [None req-f70d8e25-e795-41bd-ae53-9acba3f1fd71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315594, 'name': PowerOffVM_Task, 'duration_secs': 0.932199} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.928506] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-f70d8e25-e795-41bd-ae53-9acba3f1fd71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 897.928721] env[61868]: DEBUG nova.compute.manager [None req-f70d8e25-e795-41bd-ae53-9acba3f1fd71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 897.929601] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3746875a-56e0-4092-a7fb-6116979bfc45 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.121504] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4913bdaa-6bad-444f-8690-03eeb16ef4b9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Acquiring lock "409cad1a-946d-4c58-aa57-1c0bf97fe63f" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.125026] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4913bdaa-6bad-444f-8690-03eeb16ef4b9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Lock "409cad1a-946d-4c58-aa57-1c0bf97fe63f" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.125026] env[61868]: INFO nova.compute.manager [None req-4913bdaa-6bad-444f-8690-03eeb16ef4b9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Attaching volume 7f7cfe03-d391-4d1c-8979-7fb54c701162 to /dev/sdb [ 898.125716] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.399s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.126343] env[61868]: DEBUG nova.compute.manager [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 898.130028] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 24.511s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.130360] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.131261] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61868) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 898.135145] env[61868]: DEBUG oslo_concurrency.lockutils [None req-590c00d1-2bb3-45f8-b6e0-454b3a95e2bc tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.236s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.135145] env[61868]: DEBUG nova.objects.instance [None req-590c00d1-2bb3-45f8-b6e0-454b3a95e2bc tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Lazy-loading 'resources' on Instance uuid 2c00210e-3b6a-4041-9e8e-7f40ef23d716 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 898.135381] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd5b02d0-756c-4dae-9fe0-3efed3a6a8d5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.152480] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbc1fbad-3d68-4d8d-ae4c-bf310f1ed566 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.171667] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fc02afe-c4dc-4080-8019-411e125bdcb9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.176694] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6074ef6-2110-4b55-8fac-52ed23158dda {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.185962] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea22a3f4-ef8e-47f2-b8f4-e405fe251c2c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.189677] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f9570bf-8581-4999-ac9f-7a49e496a54b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.220288] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181018MB free_disk=134GB free_vcpus=48 pci_devices=None {{(pid=61868) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 898.220521] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.224783] env[61868]: DEBUG nova.virt.block_device [None req-4913bdaa-6bad-444f-8690-03eeb16ef4b9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Updating existing volume attachment record: 4b106ae8-cafa-47cb-9af8-aae8333e7c46 {{(pid=61868) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 898.248856] env[61868]: DEBUG oslo_vmware.api [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]521724df-44d6-4479-4f73-022c4ddb9b91, 'name': SearchDatastore_Task, 'duration_secs': 0.009553} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.249237] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.249537] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 898.249890] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.250139] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.250412] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 898.250739] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-21bf9902-c539-401f-a59b-876b2a41dcac {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.259478] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 898.262157] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 898.262157] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-753b4246-05ee-4b0c-a276-139a1d2048d2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.266984] env[61868]: DEBUG oslo_vmware.api [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the task: (returnval){ [ 898.266984] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]520be38e-834d-11dd-26ce-f2d43d5e3ce1" [ 898.266984] env[61868]: _type = "Task" [ 898.266984] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.282662] env[61868]: DEBUG oslo_vmware.api [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]520be38e-834d-11dd-26ce-f2d43d5e3ce1, 'name': SearchDatastore_Task, 'duration_secs': 0.009754} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.286584] env[61868]: DEBUG oslo_vmware.api [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': task-1315593, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.363438} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.286887] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34e52323-d683-4b0c-a03d-c4ca2e838c7c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.289449] env[61868]: INFO nova.virt.vmwareapi.ds_util [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d-rescue.vmdk. [ 898.290406] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d62206d-0b06-4bf7-a16f-a9a97c0aa5f1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.297074] env[61868]: DEBUG oslo_vmware.api [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the task: (returnval){ [ 898.297074] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5240516a-e430-6465-dbf7-0761c4a39b32" [ 898.297074] env[61868]: _type = "Task" [ 898.297074] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.320881] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d-rescue.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 898.325640] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2cbfe15b-f57d-407a-a50b-b14b7be01a85 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.340084] env[61868]: DEBUG nova.compute.manager [req-30afb072-08f7-4ae6-bab3-6302dcbb82f6 req-8901fe5e-de0e-4003-a7ad-f30bcbf7f902 service nova] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Received event network-vif-plugged-c27d3e06-cf3c-4ef8-ba0b-e6f2ca824646 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 898.340331] env[61868]: DEBUG oslo_concurrency.lockutils [req-30afb072-08f7-4ae6-bab3-6302dcbb82f6 req-8901fe5e-de0e-4003-a7ad-f30bcbf7f902 service nova] Acquiring lock "8ea8e28b-3582-45f2-b4b5-84f624415a58-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.341324] env[61868]: DEBUG oslo_concurrency.lockutils [req-30afb072-08f7-4ae6-bab3-6302dcbb82f6 req-8901fe5e-de0e-4003-a7ad-f30bcbf7f902 service nova] Lock "8ea8e28b-3582-45f2-b4b5-84f624415a58-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.341324] env[61868]: DEBUG oslo_concurrency.lockutils [req-30afb072-08f7-4ae6-bab3-6302dcbb82f6 req-8901fe5e-de0e-4003-a7ad-f30bcbf7f902 service nova] Lock "8ea8e28b-3582-45f2-b4b5-84f624415a58-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.341324] env[61868]: DEBUG nova.compute.manager [req-30afb072-08f7-4ae6-bab3-6302dcbb82f6 req-8901fe5e-de0e-4003-a7ad-f30bcbf7f902 service nova] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] No waiting events found dispatching network-vif-plugged-c27d3e06-cf3c-4ef8-ba0b-e6f2ca824646 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 898.341324] env[61868]: WARNING nova.compute.manager [req-30afb072-08f7-4ae6-bab3-6302dcbb82f6 req-8901fe5e-de0e-4003-a7ad-f30bcbf7f902 service nova] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Received unexpected event network-vif-plugged-c27d3e06-cf3c-4ef8-ba0b-e6f2ca824646 for instance with vm_state building and task_state spawning. [ 898.349966] env[61868]: DEBUG oslo_vmware.api [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5240516a-e430-6465-dbf7-0761c4a39b32, 'name': SearchDatastore_Task, 'duration_secs': 0.010226} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.351512] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.351780] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] e2832c0f-dc62-4628-87d4-e2a01819b771/e2832c0f-dc62-4628-87d4-e2a01819b771.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 898.352139] env[61868]: DEBUG oslo_vmware.api [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Waiting for the task: (returnval){ [ 898.352139] env[61868]: value = "task-1315595" [ 898.352139] env[61868]: _type = "Task" [ 898.352139] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.352329] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-864819a2-1463-44b9-98c8-b37c3490b552 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.362058] env[61868]: DEBUG oslo_vmware.api [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the task: (returnval){ [ 898.362058] env[61868]: value = "task-1315596" [ 898.362058] env[61868]: _type = "Task" [ 898.362058] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.365691] env[61868]: DEBUG oslo_vmware.api [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': task-1315595, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.374969] env[61868]: DEBUG oslo_vmware.api [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315596, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.442251] env[61868]: DEBUG nova.network.neutron [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Successfully updated port: c27d3e06-cf3c-4ef8-ba0b-e6f2ca824646 {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 898.447533] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f70d8e25-e795-41bd-ae53-9acba3f1fd71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "306f2dc4-933c-4948-8330-eda93931c4e0" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.057s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.636765] env[61868]: DEBUG nova.compute.utils [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 898.639447] env[61868]: DEBUG nova.compute.manager [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 898.640019] env[61868]: DEBUG nova.network.neutron [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 898.696703] env[61868]: DEBUG nova.policy [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e02ffdc140274733a081fd53c4acc202', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a60c3c20950e4156b64c3b4c61b9f0f2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 898.869694] env[61868]: DEBUG oslo_vmware.api [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': task-1315595, 'name': ReconfigVM_Task, 'duration_secs': 0.421298} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.876386] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Reconfigured VM instance instance-0000004a to attach disk [datastore1] f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d-rescue.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 898.877953] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2e29cd5-96d6-4925-b27b-fdbf7d4701a1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.916717] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-22eff2d3-6bb3-407f-b1d8-e7a195f662be {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.927702] env[61868]: DEBUG oslo_vmware.api [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315596, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.934297] env[61868]: DEBUG oslo_vmware.api [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Waiting for the task: (returnval){ [ 898.934297] env[61868]: value = "task-1315600" [ 898.934297] env[61868]: _type = "Task" [ 898.934297] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.944767] env[61868]: DEBUG oslo_vmware.api [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': task-1315600, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.948777] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Acquiring lock "refresh_cache-8ea8e28b-3582-45f2-b4b5-84f624415a58" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.948777] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Acquired lock "refresh_cache-8ea8e28b-3582-45f2-b4b5-84f624415a58" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.948777] env[61868]: DEBUG nova.network.neutron [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 899.037258] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f44b1c7d-391f-47a5-922d-484bca6c3e51 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.047298] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94fa0d64-08fc-4e1e-aabf-0a09a6de0ee2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.088507] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69de0f53-804e-453a-bfa5-1af71d02cc27 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.102895] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4e57e7c-1cb0-4d04-9b41-48db633bfcc6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.115398] env[61868]: DEBUG nova.compute.provider_tree [None req-590c00d1-2bb3-45f8-b6e0-454b3a95e2bc tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 899.123817] env[61868]: DEBUG nova.network.neutron [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Successfully created port: e3943fa9-423d-4168-8aa6-1651280cd127 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 899.146937] env[61868]: DEBUG nova.compute.manager [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 899.384413] env[61868]: DEBUG oslo_vmware.api [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315596, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.538654} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.384413] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] e2832c0f-dc62-4628-87d4-e2a01819b771/e2832c0f-dc62-4628-87d4-e2a01819b771.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 899.384413] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 899.384413] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c6aca38d-7ae5-428d-8533-6747aff9b9e2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.391701] env[61868]: DEBUG oslo_vmware.api [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the task: (returnval){ [ 899.391701] env[61868]: value = "task-1315601" [ 899.391701] env[61868]: _type = "Task" [ 899.391701] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.402089] env[61868]: DEBUG oslo_vmware.api [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315601, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.444083] env[61868]: DEBUG oslo_vmware.api [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': task-1315600, 'name': ReconfigVM_Task, 'duration_secs': 0.167619} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.448036] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 899.448036] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-86560060-7591-4b72-b0f4-0a8e38fe2b78 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.454663] env[61868]: DEBUG oslo_vmware.api [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Waiting for the task: (returnval){ [ 899.454663] env[61868]: value = "task-1315602" [ 899.454663] env[61868]: _type = "Task" [ 899.454663] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.464167] env[61868]: DEBUG oslo_vmware.api [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': task-1315602, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.483156] env[61868]: DEBUG nova.network.neutron [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 899.592818] env[61868]: DEBUG nova.network.neutron [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Updating instance_info_cache with network_info: [{"id": "c27d3e06-cf3c-4ef8-ba0b-e6f2ca824646", "address": "fa:16:3e:3e:8f:76", "network": {"id": "094eaf15-c932-4d4f-8736-fe5239160507", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1905020504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "7541757d43c74b93acf728aa2fb0f425", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc27d3e06-cf", "ovs_interfaceid": "c27d3e06-cf3c-4ef8-ba0b-e6f2ca824646", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.618379] env[61868]: DEBUG nova.scheduler.client.report [None req-590c00d1-2bb3-45f8-b6e0-454b3a95e2bc tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 899.902762] env[61868]: DEBUG oslo_vmware.api [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315601, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079125} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.903234] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 899.904175] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-373fc5e5-9aff-4b7f-a52e-d6fc5bea5c1f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.928781] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] e2832c0f-dc62-4628-87d4-e2a01819b771/e2832c0f-dc62-4628-87d4-e2a01819b771.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 899.929117] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d1a3c2b6-67da-4688-a52d-e9ad89d85bed {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.949969] env[61868]: DEBUG oslo_vmware.api [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the task: (returnval){ [ 899.949969] env[61868]: value = "task-1315603" [ 899.949969] env[61868]: _type = "Task" [ 899.949969] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.958544] env[61868]: DEBUG oslo_vmware.api [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315603, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.968893] env[61868]: DEBUG oslo_vmware.api [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': task-1315602, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.063659] env[61868]: DEBUG nova.compute.manager [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 900.064562] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df15e4bc-8ac1-4709-a551-45ed328d2dc0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.095843] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Releasing lock "refresh_cache-8ea8e28b-3582-45f2-b4b5-84f624415a58" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.096209] env[61868]: DEBUG nova.compute.manager [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Instance network_info: |[{"id": "c27d3e06-cf3c-4ef8-ba0b-e6f2ca824646", "address": "fa:16:3e:3e:8f:76", "network": {"id": "094eaf15-c932-4d4f-8736-fe5239160507", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1905020504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "7541757d43c74b93acf728aa2fb0f425", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc27d3e06-cf", "ovs_interfaceid": "c27d3e06-cf3c-4ef8-ba0b-e6f2ca824646", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 900.096760] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3e:8f:76', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69f65356-c85e-4b7f-ad28-7c7b5e8cf50c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c27d3e06-cf3c-4ef8-ba0b-e6f2ca824646', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 900.104374] env[61868]: DEBUG oslo.service.loopingcall [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 900.105240] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 900.105473] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-594bb3a1-b1b4-4f81-acc2-bbbfa70c413b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.123511] env[61868]: DEBUG oslo_concurrency.lockutils [None req-590c00d1-2bb3-45f8-b6e0-454b3a95e2bc tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.991s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.126930] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.330s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.127195] env[61868]: DEBUG nova.objects.instance [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Lazy-loading 'resources' on Instance uuid cf063dbc-7b5a-4836-91bf-a0aa33cca6bc {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 900.128344] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 900.128344] env[61868]: value = "task-1315604" [ 900.128344] env[61868]: _type = "Task" [ 900.128344] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.138590] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315604, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.157075] env[61868]: DEBUG nova.compute.manager [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 900.183513] env[61868]: DEBUG nova.virt.hardware [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 900.184024] env[61868]: DEBUG nova.virt.hardware [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 900.184377] env[61868]: DEBUG nova.virt.hardware [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 900.185298] env[61868]: DEBUG nova.virt.hardware [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 900.185298] env[61868]: DEBUG nova.virt.hardware [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 900.185298] env[61868]: DEBUG nova.virt.hardware [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 900.185298] env[61868]: DEBUG nova.virt.hardware [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 900.185510] env[61868]: DEBUG nova.virt.hardware [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 900.185726] env[61868]: DEBUG nova.virt.hardware [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 900.186545] env[61868]: DEBUG nova.virt.hardware [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 900.186545] env[61868]: DEBUG nova.virt.hardware [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 900.187166] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbf9b702-0b31-4896-b861-7c8d955e8ea4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.196031] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9c630d6-34d0-4527-845a-2277ec8888dd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.344288] env[61868]: DEBUG nova.compute.manager [req-c4e21169-75fa-4a78-81fd-869c8a67c7d2 req-2efa0608-893f-412d-a513-3936ef59df8f service nova] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Received event network-changed-c27d3e06-cf3c-4ef8-ba0b-e6f2ca824646 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 900.344495] env[61868]: DEBUG nova.compute.manager [req-c4e21169-75fa-4a78-81fd-869c8a67c7d2 req-2efa0608-893f-412d-a513-3936ef59df8f service nova] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Refreshing instance network info cache due to event network-changed-c27d3e06-cf3c-4ef8-ba0b-e6f2ca824646. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 900.344805] env[61868]: DEBUG oslo_concurrency.lockutils [req-c4e21169-75fa-4a78-81fd-869c8a67c7d2 req-2efa0608-893f-412d-a513-3936ef59df8f service nova] Acquiring lock "refresh_cache-8ea8e28b-3582-45f2-b4b5-84f624415a58" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.344881] env[61868]: DEBUG oslo_concurrency.lockutils [req-c4e21169-75fa-4a78-81fd-869c8a67c7d2 req-2efa0608-893f-412d-a513-3936ef59df8f service nova] Acquired lock "refresh_cache-8ea8e28b-3582-45f2-b4b5-84f624415a58" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.345093] env[61868]: DEBUG nova.network.neutron [req-c4e21169-75fa-4a78-81fd-869c8a67c7d2 req-2efa0608-893f-412d-a513-3936ef59df8f service nova] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Refreshing network info cache for port c27d3e06-cf3c-4ef8-ba0b-e6f2ca824646 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 900.460828] env[61868]: DEBUG oslo_vmware.api [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315603, 'name': ReconfigVM_Task, 'duration_secs': 0.321371} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.464368] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Reconfigured VM instance instance-0000004c to attach disk [datastore1] e2832c0f-dc62-4628-87d4-e2a01819b771/e2832c0f-dc62-4628-87d4-e2a01819b771.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 900.465029] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-40a0fd92-2f62-4f36-9ff8-e5c58b760d7b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.473771] env[61868]: DEBUG oslo_vmware.api [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': task-1315602, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.475537] env[61868]: DEBUG oslo_vmware.api [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the task: (returnval){ [ 900.475537] env[61868]: value = "task-1315605" [ 900.475537] env[61868]: _type = "Task" [ 900.475537] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.485750] env[61868]: DEBUG oslo_vmware.api [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315605, 'name': Rename_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.575848] env[61868]: INFO nova.compute.manager [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] instance snapshotting [ 900.576065] env[61868]: WARNING nova.compute.manager [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] trying to snapshot a non-running instance: (state: 4 expected: 1) [ 900.578767] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97ceb142-896b-41b6-8a1b-554392f29b92 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.599825] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f8a7e69-75f9-408d-9e7b-6370dff0b157 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.641979] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315604, 'name': CreateVM_Task, 'duration_secs': 0.351307} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.642206] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 900.642957] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.643173] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.643503] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 900.646324] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-28495f71-a731-49ff-b56c-14da835cc0bb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.648495] env[61868]: DEBUG oslo_concurrency.lockutils [None req-590c00d1-2bb3-45f8-b6e0-454b3a95e2bc tempest-ServerActionsV293TestJSON-1902292359 tempest-ServerActionsV293TestJSON-1902292359-project-member] Lock "2c00210e-3b6a-4041-9e8e-7f40ef23d716" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.587s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.656022] env[61868]: DEBUG oslo_vmware.api [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 900.656022] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5207b988-f431-54b2-b6f6-9bac11ca29f5" [ 900.656022] env[61868]: _type = "Task" [ 900.656022] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.669665] env[61868]: DEBUG oslo_vmware.api [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5207b988-f431-54b2-b6f6-9bac11ca29f5, 'name': SearchDatastore_Task, 'duration_secs': 0.011121} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.669995] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.670250] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 900.670484] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.670635] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.670818] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 900.671099] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-46e0f77b-d6f3-4d70-9b81-7bfb7b8f50e2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.690967] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 900.691238] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 900.692113] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-19db8e1b-f5bf-49fe-a25f-f726c89cb666 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.698346] env[61868]: DEBUG oslo_vmware.api [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 900.698346] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]522f5aba-fe86-1848-ac3c-a5461ffa94ee" [ 900.698346] env[61868]: _type = "Task" [ 900.698346] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.712778] env[61868]: DEBUG oslo_vmware.api [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]522f5aba-fe86-1848-ac3c-a5461ffa94ee, 'name': SearchDatastore_Task, 'duration_secs': 0.009841} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.716311] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b1256d3-2bc8-4060-bcda-455d7932d03e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.722321] env[61868]: DEBUG oslo_vmware.api [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 900.722321] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]520acb0a-0cb8-71f7-55af-9bf5cac329a9" [ 900.722321] env[61868]: _type = "Task" [ 900.722321] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.735243] env[61868]: DEBUG oslo_vmware.api [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]520acb0a-0cb8-71f7-55af-9bf5cac329a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.785069] env[61868]: DEBUG nova.network.neutron [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Successfully updated port: e3943fa9-423d-4168-8aa6-1651280cd127 {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 900.962336] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0fb076e-3640-497f-9264-93d77697f2d2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.975980] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df7c7af5-22b3-4c38-905d-191affbcf13f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.979110] env[61868]: DEBUG oslo_vmware.api [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': task-1315602, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.989421] env[61868]: DEBUG oslo_vmware.api [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315605, 'name': Rename_Task, 'duration_secs': 0.150658} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.015041] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 901.017966] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c5b18c3d-8bb9-4acf-b3f7-2b5aa80718bf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.020184] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baec2793-0718-4826-9224-406e7cbdbd65 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.028932] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea24b381-f6e6-4915-9a7b-4fb1e43487ea {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.034199] env[61868]: DEBUG oslo_vmware.api [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the task: (returnval){ [ 901.034199] env[61868]: value = "task-1315607" [ 901.034199] env[61868]: _type = "Task" [ 901.034199] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.048589] env[61868]: DEBUG nova.compute.provider_tree [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 901.056941] env[61868]: DEBUG oslo_vmware.api [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315607, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.111756] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Creating Snapshot of the VM instance {{(pid=61868) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 901.112168] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-e8527b73-49bf-47ed-9866-6298d428e26e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.123965] env[61868]: DEBUG oslo_vmware.api [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 901.123965] env[61868]: value = "task-1315608" [ 901.123965] env[61868]: _type = "Task" [ 901.123965] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.132564] env[61868]: DEBUG oslo_vmware.api [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315608, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.161169] env[61868]: DEBUG nova.network.neutron [req-c4e21169-75fa-4a78-81fd-869c8a67c7d2 req-2efa0608-893f-412d-a513-3936ef59df8f service nova] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Updated VIF entry in instance network info cache for port c27d3e06-cf3c-4ef8-ba0b-e6f2ca824646. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 901.161740] env[61868]: DEBUG nova.network.neutron [req-c4e21169-75fa-4a78-81fd-869c8a67c7d2 req-2efa0608-893f-412d-a513-3936ef59df8f service nova] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Updating instance_info_cache with network_info: [{"id": "c27d3e06-cf3c-4ef8-ba0b-e6f2ca824646", "address": "fa:16:3e:3e:8f:76", "network": {"id": "094eaf15-c932-4d4f-8736-fe5239160507", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1905020504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "7541757d43c74b93acf728aa2fb0f425", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc27d3e06-cf", "ovs_interfaceid": "c27d3e06-cf3c-4ef8-ba0b-e6f2ca824646", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.233723] env[61868]: DEBUG oslo_vmware.api [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]520acb0a-0cb8-71f7-55af-9bf5cac329a9, 'name': SearchDatastore_Task, 'duration_secs': 0.012465} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.234058] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.234288] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 8ea8e28b-3582-45f2-b4b5-84f624415a58/8ea8e28b-3582-45f2-b4b5-84f624415a58.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 901.234560] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7d42b459-1262-48e7-bf73-cf32f2000f59 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.242243] env[61868]: DEBUG oslo_vmware.api [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 901.242243] env[61868]: value = "task-1315609" [ 901.242243] env[61868]: _type = "Task" [ 901.242243] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.250707] env[61868]: DEBUG oslo_vmware.api [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315609, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.287933] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "refresh_cache-842c8606-632a-47d4-9a67-38d8e6ce6b4d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.288239] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquired lock "refresh_cache-842c8606-632a-47d4-9a67-38d8e6ce6b4d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.288239] env[61868]: DEBUG nova.network.neutron [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 901.473137] env[61868]: DEBUG oslo_vmware.api [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': task-1315602, 'name': PowerOnVM_Task, 'duration_secs': 1.601256} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.473570] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 901.476972] env[61868]: DEBUG nova.compute.manager [None req-6502ee16-cafd-474f-8f4b-57e42979b4e1 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 901.477939] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-519a7220-2332-4d3d-817c-6f0b7421aad2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.546271] env[61868]: DEBUG oslo_vmware.api [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315607, 'name': PowerOnVM_Task} progress is 90%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.555023] env[61868]: DEBUG nova.scheduler.client.report [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 901.636251] env[61868]: DEBUG oslo_vmware.api [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315608, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.664673] env[61868]: DEBUG oslo_concurrency.lockutils [req-c4e21169-75fa-4a78-81fd-869c8a67c7d2 req-2efa0608-893f-412d-a513-3936ef59df8f service nova] Releasing lock "refresh_cache-8ea8e28b-3582-45f2-b4b5-84f624415a58" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.753102] env[61868]: DEBUG oslo_vmware.api [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315609, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.826508] env[61868]: DEBUG nova.network.neutron [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 902.045198] env[61868]: DEBUG oslo_vmware.api [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315607, 'name': PowerOnVM_Task, 'duration_secs': 0.720676} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.046093] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 902.046312] env[61868]: INFO nova.compute.manager [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Took 8.66 seconds to spawn the instance on the hypervisor. [ 902.046497] env[61868]: DEBUG nova.compute.manager [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 902.047288] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24303abf-7b5e-4629-9dcf-77e8e40dc35d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.059774] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.933s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.062988] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9d9434da-029e-4450-831a-50e8cce87561 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.241s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.063256] env[61868]: DEBUG nova.objects.instance [None req-9d9434da-029e-4450-831a-50e8cce87561 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lazy-loading 'resources' on Instance uuid c7b10943-5136-44c8-b4e7-59651220a333 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 902.088261] env[61868]: INFO nova.scheduler.client.report [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Deleted allocations for instance cf063dbc-7b5a-4836-91bf-a0aa33cca6bc [ 902.135224] env[61868]: DEBUG oslo_vmware.api [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315608, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.233212] env[61868]: DEBUG nova.network.neutron [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Updating instance_info_cache with network_info: [{"id": "e3943fa9-423d-4168-8aa6-1651280cd127", "address": "fa:16:3e:1d:68:a3", "network": {"id": "c1f5bf03-438f-4acc-93fd-159bed7d4069", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2061742146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a60c3c20950e4156b64c3b4c61b9f0f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24b91e3b-50f1-4a16-b929-942e6b31b2d8", "external-id": "nsx-vlan-transportzone-34", "segmentation_id": 34, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3943fa9-42", "ovs_interfaceid": "e3943fa9-423d-4168-8aa6-1651280cd127", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.256166] env[61868]: DEBUG oslo_vmware.api [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315609, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.522048} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.256445] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 8ea8e28b-3582-45f2-b4b5-84f624415a58/8ea8e28b-3582-45f2-b4b5-84f624415a58.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 902.256655] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 902.256903] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d5ed382d-8373-488d-b085-0c36fb756ce0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.265489] env[61868]: DEBUG oslo_vmware.api [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 902.265489] env[61868]: value = "task-1315610" [ 902.265489] env[61868]: _type = "Task" [ 902.265489] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.422172] env[61868]: DEBUG nova.compute.manager [req-d0923c39-7a76-4cfd-9935-e2344e4c46bb req-31e62f79-6252-4199-8834-a588fd3e570f service nova] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Received event network-vif-plugged-e3943fa9-423d-4168-8aa6-1651280cd127 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 902.422400] env[61868]: DEBUG oslo_concurrency.lockutils [req-d0923c39-7a76-4cfd-9935-e2344e4c46bb req-31e62f79-6252-4199-8834-a588fd3e570f service nova] Acquiring lock "842c8606-632a-47d4-9a67-38d8e6ce6b4d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.422615] env[61868]: DEBUG oslo_concurrency.lockutils [req-d0923c39-7a76-4cfd-9935-e2344e4c46bb req-31e62f79-6252-4199-8834-a588fd3e570f service nova] Lock "842c8606-632a-47d4-9a67-38d8e6ce6b4d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.422787] env[61868]: DEBUG oslo_concurrency.lockutils [req-d0923c39-7a76-4cfd-9935-e2344e4c46bb req-31e62f79-6252-4199-8834-a588fd3e570f service nova] Lock "842c8606-632a-47d4-9a67-38d8e6ce6b4d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.422983] env[61868]: DEBUG nova.compute.manager [req-d0923c39-7a76-4cfd-9935-e2344e4c46bb req-31e62f79-6252-4199-8834-a588fd3e570f service nova] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] No waiting events found dispatching network-vif-plugged-e3943fa9-423d-4168-8aa6-1651280cd127 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 902.423192] env[61868]: WARNING nova.compute.manager [req-d0923c39-7a76-4cfd-9935-e2344e4c46bb req-31e62f79-6252-4199-8834-a588fd3e570f service nova] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Received unexpected event network-vif-plugged-e3943fa9-423d-4168-8aa6-1651280cd127 for instance with vm_state building and task_state spawning. [ 902.423357] env[61868]: DEBUG nova.compute.manager [req-d0923c39-7a76-4cfd-9935-e2344e4c46bb req-31e62f79-6252-4199-8834-a588fd3e570f service nova] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Received event network-changed-e3943fa9-423d-4168-8aa6-1651280cd127 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 902.423512] env[61868]: DEBUG nova.compute.manager [req-d0923c39-7a76-4cfd-9935-e2344e4c46bb req-31e62f79-6252-4199-8834-a588fd3e570f service nova] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Refreshing instance network info cache due to event network-changed-e3943fa9-423d-4168-8aa6-1651280cd127. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 902.423678] env[61868]: DEBUG oslo_concurrency.lockutils [req-d0923c39-7a76-4cfd-9935-e2344e4c46bb req-31e62f79-6252-4199-8834-a588fd3e570f service nova] Acquiring lock "refresh_cache-842c8606-632a-47d4-9a67-38d8e6ce6b4d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.570983] env[61868]: INFO nova.compute.manager [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Took 32.56 seconds to build instance. [ 902.596202] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b1b55f62-4d33-4d32-86dc-b0c49b88f3c4 tempest-ServersAaction247Test-1620107749 tempest-ServersAaction247Test-1620107749-project-member] Lock "cf063dbc-7b5a-4836-91bf-a0aa33cca6bc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.215s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.635715] env[61868]: DEBUG oslo_vmware.api [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315608, 'name': CreateSnapshot_Task, 'duration_secs': 1.071953} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.636418] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Created Snapshot of the VM instance {{(pid=61868) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 902.637877] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f422022-e98c-43d6-82de-3189916a0ab3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.736344] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Releasing lock "refresh_cache-842c8606-632a-47d4-9a67-38d8e6ce6b4d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.736685] env[61868]: DEBUG nova.compute.manager [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Instance network_info: |[{"id": "e3943fa9-423d-4168-8aa6-1651280cd127", "address": "fa:16:3e:1d:68:a3", "network": {"id": "c1f5bf03-438f-4acc-93fd-159bed7d4069", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2061742146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a60c3c20950e4156b64c3b4c61b9f0f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24b91e3b-50f1-4a16-b929-942e6b31b2d8", "external-id": "nsx-vlan-transportzone-34", "segmentation_id": 34, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3943fa9-42", "ovs_interfaceid": "e3943fa9-423d-4168-8aa6-1651280cd127", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 902.739360] env[61868]: DEBUG oslo_concurrency.lockutils [req-d0923c39-7a76-4cfd-9935-e2344e4c46bb req-31e62f79-6252-4199-8834-a588fd3e570f service nova] Acquired lock "refresh_cache-842c8606-632a-47d4-9a67-38d8e6ce6b4d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.739360] env[61868]: DEBUG nova.network.neutron [req-d0923c39-7a76-4cfd-9935-e2344e4c46bb req-31e62f79-6252-4199-8834-a588fd3e570f service nova] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Refreshing network info cache for port e3943fa9-423d-4168-8aa6-1651280cd127 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 902.739360] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1d:68:a3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24b91e3b-50f1-4a16-b929-942e6b31b2d8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e3943fa9-423d-4168-8aa6-1651280cd127', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 902.745958] env[61868]: DEBUG oslo.service.loopingcall [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 902.749542] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 902.749956] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6f91c363-f5bd-4118-b6bc-975e4494b3ec {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.777297] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 902.777297] env[61868]: value = "task-1315611" [ 902.777297] env[61868]: _type = "Task" [ 902.777297] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.788109] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-4913bdaa-6bad-444f-8690-03eeb16ef4b9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Volume attach. Driver type: vmdk {{(pid=61868) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 902.788367] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-4913bdaa-6bad-444f-8690-03eeb16ef4b9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281562', 'volume_id': '7f7cfe03-d391-4d1c-8979-7fb54c701162', 'name': 'volume-7f7cfe03-d391-4d1c-8979-7fb54c701162', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '409cad1a-946d-4c58-aa57-1c0bf97fe63f', 'attached_at': '', 'detached_at': '', 'volume_id': '7f7cfe03-d391-4d1c-8979-7fb54c701162', 'serial': '7f7cfe03-d391-4d1c-8979-7fb54c701162'} {{(pid=61868) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 902.792090] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d325c9b3-1b10-4a1c-831a-8d79d275874f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.794980] env[61868]: DEBUG oslo_vmware.api [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315610, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088019} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.795876] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 902.796950] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0fc676f-4d98-4e4f-a8fc-dd0236d37bee {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.815140] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315611, 'name': CreateVM_Task} progress is 10%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.816649] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64ce0cbc-5bb9-4c14-be6b-3c371be135d6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.838038] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] 8ea8e28b-3582-45f2-b4b5-84f624415a58/8ea8e28b-3582-45f2-b4b5-84f624415a58.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 902.841230] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c70d86e-1e3b-48be-bd2c-99d6c5543b3a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.878089] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-4913bdaa-6bad-444f-8690-03eeb16ef4b9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] volume-7f7cfe03-d391-4d1c-8979-7fb54c701162/volume-7f7cfe03-d391-4d1c-8979-7fb54c701162.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 902.878905] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1db23d56-a773-4f16-8336-4ffe1380104b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.895376] env[61868]: DEBUG oslo_vmware.api [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 902.895376] env[61868]: value = "task-1315612" [ 902.895376] env[61868]: _type = "Task" [ 902.895376] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.901579] env[61868]: DEBUG oslo_vmware.api [None req-4913bdaa-6bad-444f-8690-03eeb16ef4b9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Waiting for the task: (returnval){ [ 902.901579] env[61868]: value = "task-1315613" [ 902.901579] env[61868]: _type = "Task" [ 902.901579] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.908743] env[61868]: DEBUG oslo_vmware.api [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315612, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.917278] env[61868]: DEBUG oslo_vmware.api [None req-4913bdaa-6bad-444f-8690-03eeb16ef4b9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315613, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.974176] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0851172-0cfe-4d08-a1d3-4ef989818d13 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.983227] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-844af987-d3fa-44d6-8e84-089bf6f72b82 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.015154] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62f75881-0ccd-43a9-b0c8-25de1fb04c2a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.023866] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68cb16fa-64b8-4969-b216-68b80ff35732 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.039621] env[61868]: DEBUG nova.compute.provider_tree [None req-9d9434da-029e-4450-831a-50e8cce87561 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 903.073355] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7ad9d4a2-efaf-430e-9999-3788322148c9 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lock "e2832c0f-dc62-4628-87d4-e2a01819b771" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.075s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.163624] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Creating linked-clone VM from snapshot {{(pid=61868) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 903.163963] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-bd751f58-6f4b-43f7-bee5-646adffc5a4c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.173749] env[61868]: DEBUG oslo_vmware.api [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 903.173749] env[61868]: value = "task-1315614" [ 903.173749] env[61868]: _type = "Task" [ 903.173749] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.187681] env[61868]: DEBUG oslo_vmware.api [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315614, 'name': CloneVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.294024] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315611, 'name': CreateVM_Task, 'duration_secs': 0.366945} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.294024] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 903.294024] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.294024] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.294024] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 903.294024] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-838226ab-1c6c-4383-badc-cd56ef381764 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.299714] env[61868]: DEBUG oslo_vmware.api [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 903.299714] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52245f83-92f2-9fbc-fa97-a670917a7d69" [ 903.299714] env[61868]: _type = "Task" [ 903.299714] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.310404] env[61868]: DEBUG oslo_vmware.api [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52245f83-92f2-9fbc-fa97-a670917a7d69, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.409969] env[61868]: DEBUG oslo_vmware.api [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315612, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.416614] env[61868]: DEBUG oslo_vmware.api [None req-4913bdaa-6bad-444f-8690-03eeb16ef4b9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315613, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.545025] env[61868]: DEBUG nova.scheduler.client.report [None req-9d9434da-029e-4450-831a-50e8cce87561 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 903.685586] env[61868]: DEBUG oslo_vmware.api [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315614, 'name': CloneVM_Task} progress is 94%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.763622] env[61868]: DEBUG nova.network.neutron [req-d0923c39-7a76-4cfd-9935-e2344e4c46bb req-31e62f79-6252-4199-8834-a588fd3e570f service nova] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Updated VIF entry in instance network info cache for port e3943fa9-423d-4168-8aa6-1651280cd127. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 903.763622] env[61868]: DEBUG nova.network.neutron [req-d0923c39-7a76-4cfd-9935-e2344e4c46bb req-31e62f79-6252-4199-8834-a588fd3e570f service nova] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Updating instance_info_cache with network_info: [{"id": "e3943fa9-423d-4168-8aa6-1651280cd127", "address": "fa:16:3e:1d:68:a3", "network": {"id": "c1f5bf03-438f-4acc-93fd-159bed7d4069", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2061742146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a60c3c20950e4156b64c3b4c61b9f0f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24b91e3b-50f1-4a16-b929-942e6b31b2d8", "external-id": "nsx-vlan-transportzone-34", "segmentation_id": 34, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3943fa9-42", "ovs_interfaceid": "e3943fa9-423d-4168-8aa6-1651280cd127", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.812955] env[61868]: DEBUG oslo_vmware.api [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52245f83-92f2-9fbc-fa97-a670917a7d69, 'name': SearchDatastore_Task, 'duration_secs': 0.011463} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.815083] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.815083] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 903.815083] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.815083] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.815285] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 903.815285] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-26a3f293-5d5f-4a47-b1b3-5ff5451bdc21 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.829508] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 903.829508] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 903.829508] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12ec68c3-7627-426d-b657-c48637121c5d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.837024] env[61868]: DEBUG oslo_vmware.api [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 903.837024] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]522c41d3-f8ee-7fdb-5950-adab018b4634" [ 903.837024] env[61868]: _type = "Task" [ 903.837024] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.844301] env[61868]: DEBUG oslo_vmware.api [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]522c41d3-f8ee-7fdb-5950-adab018b4634, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.913706] env[61868]: DEBUG oslo_vmware.api [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315612, 'name': ReconfigVM_Task, 'duration_secs': 0.660631} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.916933] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Reconfigured VM instance instance-0000004d to attach disk [datastore1] 8ea8e28b-3582-45f2-b4b5-84f624415a58/8ea8e28b-3582-45f2-b4b5-84f624415a58.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 903.917748] env[61868]: DEBUG oslo_vmware.api [None req-4913bdaa-6bad-444f-8690-03eeb16ef4b9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315613, 'name': ReconfigVM_Task, 'duration_secs': 0.731375} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.918049] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4c31aaf8-183b-4da1-bdd6-725a4627f48e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.919884] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-4913bdaa-6bad-444f-8690-03eeb16ef4b9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Reconfigured VM instance instance-0000003f to attach disk [datastore2] volume-7f7cfe03-d391-4d1c-8979-7fb54c701162/volume-7f7cfe03-d391-4d1c-8979-7fb54c701162.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 903.931397] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-55c3b113-6774-4701-a306-03672487279b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.947525] env[61868]: DEBUG oslo_vmware.api [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 903.947525] env[61868]: value = "task-1315615" [ 903.947525] env[61868]: _type = "Task" [ 903.947525] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.948928] env[61868]: DEBUG oslo_vmware.api [None req-4913bdaa-6bad-444f-8690-03eeb16ef4b9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Waiting for the task: (returnval){ [ 903.948928] env[61868]: value = "task-1315616" [ 903.948928] env[61868]: _type = "Task" [ 903.948928] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.959082] env[61868]: DEBUG nova.network.neutron [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Successfully updated port: 79fe6cfd-0c91-4ec5-95b2-22aed62a0e0f {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 903.967450] env[61868]: DEBUG oslo_vmware.api [None req-4913bdaa-6bad-444f-8690-03eeb16ef4b9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315616, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.967977] env[61868]: DEBUG oslo_vmware.api [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315615, 'name': Rename_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.051450] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9d9434da-029e-4450-831a-50e8cce87561 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.988s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.056022] env[61868]: DEBUG oslo_concurrency.lockutils [None req-62918748-546f-4030-9046-6bb7ec988ab2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.819s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.056022] env[61868]: DEBUG nova.objects.instance [None req-62918748-546f-4030-9046-6bb7ec988ab2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Lazy-loading 'resources' on Instance uuid 52a5afc6-4e79-436c-bc94-b61ca9fb860c {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 904.081164] env[61868]: INFO nova.scheduler.client.report [None req-9d9434da-029e-4450-831a-50e8cce87561 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Deleted allocations for instance c7b10943-5136-44c8-b4e7-59651220a333 [ 904.142434] env[61868]: DEBUG nova.compute.manager [req-013b7dbd-f86c-4518-9b3d-53100865f010 req-885b5529-60ee-45ab-8667-7dd279220923 service nova] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Received event network-vif-plugged-79fe6cfd-0c91-4ec5-95b2-22aed62a0e0f {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 904.142655] env[61868]: DEBUG oslo_concurrency.lockutils [req-013b7dbd-f86c-4518-9b3d-53100865f010 req-885b5529-60ee-45ab-8667-7dd279220923 service nova] Acquiring lock "083140bf-233d-49ac-8247-cd93206f9200-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.142867] env[61868]: DEBUG oslo_concurrency.lockutils [req-013b7dbd-f86c-4518-9b3d-53100865f010 req-885b5529-60ee-45ab-8667-7dd279220923 service nova] Lock "083140bf-233d-49ac-8247-cd93206f9200-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.143078] env[61868]: DEBUG oslo_concurrency.lockutils [req-013b7dbd-f86c-4518-9b3d-53100865f010 req-885b5529-60ee-45ab-8667-7dd279220923 service nova] Lock "083140bf-233d-49ac-8247-cd93206f9200-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.143339] env[61868]: DEBUG nova.compute.manager [req-013b7dbd-f86c-4518-9b3d-53100865f010 req-885b5529-60ee-45ab-8667-7dd279220923 service nova] [instance: 083140bf-233d-49ac-8247-cd93206f9200] No waiting events found dispatching network-vif-plugged-79fe6cfd-0c91-4ec5-95b2-22aed62a0e0f {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 904.143457] env[61868]: WARNING nova.compute.manager [req-013b7dbd-f86c-4518-9b3d-53100865f010 req-885b5529-60ee-45ab-8667-7dd279220923 service nova] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Received unexpected event network-vif-plugged-79fe6cfd-0c91-4ec5-95b2-22aed62a0e0f for instance with vm_state building and task_state deleting. [ 904.189194] env[61868]: DEBUG oslo_vmware.api [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315614, 'name': CloneVM_Task} progress is 94%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.266340] env[61868]: DEBUG oslo_concurrency.lockutils [req-d0923c39-7a76-4cfd-9935-e2344e4c46bb req-31e62f79-6252-4199-8834-a588fd3e570f service nova] Releasing lock "refresh_cache-842c8606-632a-47d4-9a67-38d8e6ce6b4d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.348877] env[61868]: DEBUG oslo_vmware.api [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]522c41d3-f8ee-7fdb-5950-adab018b4634, 'name': SearchDatastore_Task, 'duration_secs': 0.015399} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.350332] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24c2430c-bdf1-485a-a9e5-a5297b071933 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.357237] env[61868]: DEBUG oslo_vmware.api [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 904.357237] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52c583ee-8b33-090a-504e-bd08f971fbb4" [ 904.357237] env[61868]: _type = "Task" [ 904.357237] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.367906] env[61868]: DEBUG oslo_vmware.api [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52c583ee-8b33-090a-504e-bd08f971fbb4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.463547] env[61868]: DEBUG oslo_vmware.api [None req-4913bdaa-6bad-444f-8690-03eeb16ef4b9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315616, 'name': ReconfigVM_Task, 'duration_secs': 0.171903} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.466671] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-4913bdaa-6bad-444f-8690-03eeb16ef4b9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281562', 'volume_id': '7f7cfe03-d391-4d1c-8979-7fb54c701162', 'name': 'volume-7f7cfe03-d391-4d1c-8979-7fb54c701162', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '409cad1a-946d-4c58-aa57-1c0bf97fe63f', 'attached_at': '', 'detached_at': '', 'volume_id': '7f7cfe03-d391-4d1c-8979-7fb54c701162', 'serial': '7f7cfe03-d391-4d1c-8979-7fb54c701162'} {{(pid=61868) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 904.469730] env[61868]: DEBUG oslo_vmware.api [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315615, 'name': Rename_Task, 'duration_secs': 0.184077} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.469730] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 904.469730] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d3c2d58d-ba76-44b3-99e4-d32fd9849c64 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.471238] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Acquiring lock "refresh_cache-083140bf-233d-49ac-8247-cd93206f9200" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.471238] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Acquired lock "refresh_cache-083140bf-233d-49ac-8247-cd93206f9200" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.471360] env[61868]: DEBUG nova.network.neutron [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 904.481936] env[61868]: DEBUG oslo_concurrency.lockutils [None req-780bfa57-3788-4c6b-93f4-f83d47616fd0 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquiring lock "e2832c0f-dc62-4628-87d4-e2a01819b771" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.482392] env[61868]: DEBUG oslo_concurrency.lockutils [None req-780bfa57-3788-4c6b-93f4-f83d47616fd0 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lock "e2832c0f-dc62-4628-87d4-e2a01819b771" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.003s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.482664] env[61868]: DEBUG oslo_concurrency.lockutils [None req-780bfa57-3788-4c6b-93f4-f83d47616fd0 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquiring lock "e2832c0f-dc62-4628-87d4-e2a01819b771-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.482887] env[61868]: DEBUG oslo_concurrency.lockutils [None req-780bfa57-3788-4c6b-93f4-f83d47616fd0 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lock "e2832c0f-dc62-4628-87d4-e2a01819b771-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.483241] env[61868]: DEBUG oslo_concurrency.lockutils [None req-780bfa57-3788-4c6b-93f4-f83d47616fd0 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lock "e2832c0f-dc62-4628-87d4-e2a01819b771-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.484861] env[61868]: DEBUG oslo_vmware.api [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 904.484861] env[61868]: value = "task-1315617" [ 904.484861] env[61868]: _type = "Task" [ 904.484861] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.485874] env[61868]: INFO nova.compute.manager [None req-780bfa57-3788-4c6b-93f4-f83d47616fd0 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Terminating instance [ 904.490031] env[61868]: DEBUG nova.compute.manager [None req-780bfa57-3788-4c6b-93f4-f83d47616fd0 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 904.490602] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-780bfa57-3788-4c6b-93f4-f83d47616fd0 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 904.495429] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8da591d4-b7c4-4a1c-b112-1de63af2525d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.506780] env[61868]: DEBUG oslo_vmware.api [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315617, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.510074] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-780bfa57-3788-4c6b-93f4-f83d47616fd0 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 904.510562] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e3ab5fa6-2f0c-407f-b7b6-1be19c501c5d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.518665] env[61868]: DEBUG oslo_vmware.api [None req-780bfa57-3788-4c6b-93f4-f83d47616fd0 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the task: (returnval){ [ 904.518665] env[61868]: value = "task-1315618" [ 904.518665] env[61868]: _type = "Task" [ 904.518665] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.528530] env[61868]: DEBUG oslo_vmware.api [None req-780bfa57-3788-4c6b-93f4-f83d47616fd0 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315618, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.592942] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9d9434da-029e-4450-831a-50e8cce87561 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "c7b10943-5136-44c8-b4e7-59651220a333" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.193s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.596142] env[61868]: DEBUG nova.compute.manager [req-1de7ed7e-6052-4271-85fb-251feed25bd0 req-52f64ddb-8448-4899-a022-893ff7391b77 service nova] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Received event network-changed-29e8f7bd-db77-4071-b4c6-79e8338c2e82 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 904.596808] env[61868]: DEBUG nova.compute.manager [req-1de7ed7e-6052-4271-85fb-251feed25bd0 req-52f64ddb-8448-4899-a022-893ff7391b77 service nova] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Refreshing instance network info cache due to event network-changed-29e8f7bd-db77-4071-b4c6-79e8338c2e82. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 904.596808] env[61868]: DEBUG oslo_concurrency.lockutils [req-1de7ed7e-6052-4271-85fb-251feed25bd0 req-52f64ddb-8448-4899-a022-893ff7391b77 service nova] Acquiring lock "refresh_cache-f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.596808] env[61868]: DEBUG oslo_concurrency.lockutils [req-1de7ed7e-6052-4271-85fb-251feed25bd0 req-52f64ddb-8448-4899-a022-893ff7391b77 service nova] Acquired lock "refresh_cache-f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.597061] env[61868]: DEBUG nova.network.neutron [req-1de7ed7e-6052-4271-85fb-251feed25bd0 req-52f64ddb-8448-4899-a022-893ff7391b77 service nova] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Refreshing network info cache for port 29e8f7bd-db77-4071-b4c6-79e8338c2e82 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 904.691309] env[61868]: DEBUG oslo_vmware.api [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315614, 'name': CloneVM_Task} progress is 100%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.858962] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9260ad6-2395-4eaf-8be5-da2b230ad145 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.870980] env[61868]: DEBUG oslo_vmware.api [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52c583ee-8b33-090a-504e-bd08f971fbb4, 'name': SearchDatastore_Task, 'duration_secs': 0.011925} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.872981] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.873304] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 842c8606-632a-47d4-9a67-38d8e6ce6b4d/842c8606-632a-47d4-9a67-38d8e6ce6b4d.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 904.873622] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8b6a5cad-2649-4a5c-af81-25fa0ecaeb24 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.876217] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a95b73f-be22-4f40-a0a7-2c4e4a080fe5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.914077] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a70a518-9aac-4283-8df7-d3cf0fd1db7e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.916977] env[61868]: DEBUG oslo_vmware.api [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 904.916977] env[61868]: value = "task-1315619" [ 904.916977] env[61868]: _type = "Task" [ 904.916977] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.924448] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fd1141b-c8ea-4f4e-a6f3-21939b9a3d2e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.931684] env[61868]: DEBUG oslo_vmware.api [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315619, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.943078] env[61868]: DEBUG nova.compute.provider_tree [None req-62918748-546f-4030-9046-6bb7ec988ab2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 904.999507] env[61868]: DEBUG oslo_vmware.api [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315617, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.018195] env[61868]: DEBUG nova.network.neutron [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 905.029156] env[61868]: DEBUG oslo_vmware.api [None req-780bfa57-3788-4c6b-93f4-f83d47616fd0 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315618, 'name': PowerOffVM_Task, 'duration_secs': 0.262687} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.030035] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-780bfa57-3788-4c6b-93f4-f83d47616fd0 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 905.030209] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-780bfa57-3788-4c6b-93f4-f83d47616fd0 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 905.030467] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2060232c-51e9-4313-ae13-f77423337011 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.105088] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-780bfa57-3788-4c6b-93f4-f83d47616fd0 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 905.105493] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-780bfa57-3788-4c6b-93f4-f83d47616fd0 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 905.105597] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-780bfa57-3788-4c6b-93f4-f83d47616fd0 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Deleting the datastore file [datastore1] e2832c0f-dc62-4628-87d4-e2a01819b771 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 905.110020] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eb4b6b67-4ae7-40f6-a9bd-c4714910df4a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.116723] env[61868]: DEBUG oslo_vmware.api [None req-780bfa57-3788-4c6b-93f4-f83d47616fd0 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the task: (returnval){ [ 905.116723] env[61868]: value = "task-1315621" [ 905.116723] env[61868]: _type = "Task" [ 905.116723] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.126761] env[61868]: DEBUG oslo_vmware.api [None req-780bfa57-3788-4c6b-93f4-f83d47616fd0 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315621, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.189974] env[61868]: DEBUG oslo_vmware.api [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315614, 'name': CloneVM_Task, 'duration_secs': 1.579128} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.190157] env[61868]: INFO nova.virt.vmwareapi.vmops [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Created linked-clone VM from snapshot [ 905.190955] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bb73049-49b1-4169-a8e6-ca7291d5f7db {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.205153] env[61868]: DEBUG nova.virt.vmwareapi.images [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Uploading image ba384da7-fddc-4e13-bac6-b4e3d1857675 {{(pid=61868) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 905.231632] env[61868]: DEBUG oslo_vmware.rw_handles [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 905.231632] env[61868]: value = "vm-281566" [ 905.231632] env[61868]: _type = "VirtualMachine" [ 905.231632] env[61868]: }. {{(pid=61868) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 905.232036] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-8b896e1c-2676-4a8e-abf1-4071af89ff3d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.241921] env[61868]: DEBUG oslo_vmware.rw_handles [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lease: (returnval){ [ 905.241921] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52c89c03-64c0-c0ad-ba63-37391b53a30c" [ 905.241921] env[61868]: _type = "HttpNfcLease" [ 905.241921] env[61868]: } obtained for exporting VM: (result){ [ 905.241921] env[61868]: value = "vm-281566" [ 905.241921] env[61868]: _type = "VirtualMachine" [ 905.241921] env[61868]: }. {{(pid=61868) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 905.243067] env[61868]: DEBUG oslo_vmware.api [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the lease: (returnval){ [ 905.243067] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52c89c03-64c0-c0ad-ba63-37391b53a30c" [ 905.243067] env[61868]: _type = "HttpNfcLease" [ 905.243067] env[61868]: } to be ready. {{(pid=61868) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 905.251272] env[61868]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 905.251272] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52c89c03-64c0-c0ad-ba63-37391b53a30c" [ 905.251272] env[61868]: _type = "HttpNfcLease" [ 905.251272] env[61868]: } is initializing. {{(pid=61868) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 905.278133] env[61868]: DEBUG nova.network.neutron [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Updating instance_info_cache with network_info: [{"id": "79fe6cfd-0c91-4ec5-95b2-22aed62a0e0f", "address": "fa:16:3e:b7:c9:de", "network": {"id": "adc6ef1d-8fd6-4116-a2ff-34644d243811", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2090367503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f09bf553bf8b47e6846c59523d986edc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79fe6cfd-0c", "ovs_interfaceid": "79fe6cfd-0c91-4ec5-95b2-22aed62a0e0f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.427770] env[61868]: DEBUG oslo_vmware.api [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315619, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.445861] env[61868]: DEBUG nova.scheduler.client.report [None req-62918748-546f-4030-9046-6bb7ec988ab2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 905.501059] env[61868]: DEBUG oslo_vmware.api [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315617, 'name': PowerOnVM_Task, 'duration_secs': 0.783898} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.501566] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 905.501902] env[61868]: INFO nova.compute.manager [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Took 7.75 seconds to spawn the instance on the hypervisor. [ 905.502075] env[61868]: DEBUG nova.compute.manager [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 905.503113] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8632187a-c494-4399-8a81-5b997d64b4c5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.516809] env[61868]: DEBUG nova.objects.instance [None req-4913bdaa-6bad-444f-8690-03eeb16ef4b9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Lazy-loading 'flavor' on Instance uuid 409cad1a-946d-4c58-aa57-1c0bf97fe63f {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 905.631074] env[61868]: DEBUG oslo_vmware.api [None req-780bfa57-3788-4c6b-93f4-f83d47616fd0 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315621, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.35476} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.631367] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-780bfa57-3788-4c6b-93f4-f83d47616fd0 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 905.631557] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-780bfa57-3788-4c6b-93f4-f83d47616fd0 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 905.631822] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-780bfa57-3788-4c6b-93f4-f83d47616fd0 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 905.632065] env[61868]: INFO nova.compute.manager [None req-780bfa57-3788-4c6b-93f4-f83d47616fd0 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Took 1.14 seconds to destroy the instance on the hypervisor. [ 905.632324] env[61868]: DEBUG oslo.service.loopingcall [None req-780bfa57-3788-4c6b-93f4-f83d47616fd0 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 905.632520] env[61868]: DEBUG nova.compute.manager [-] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 905.632614] env[61868]: DEBUG nova.network.neutron [-] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 905.754658] env[61868]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 905.754658] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52c89c03-64c0-c0ad-ba63-37391b53a30c" [ 905.754658] env[61868]: _type = "HttpNfcLease" [ 905.754658] env[61868]: } is ready. {{(pid=61868) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 905.755037] env[61868]: DEBUG oslo_vmware.rw_handles [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 905.755037] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52c89c03-64c0-c0ad-ba63-37391b53a30c" [ 905.755037] env[61868]: _type = "HttpNfcLease" [ 905.755037] env[61868]: }. {{(pid=61868) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 905.756698] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07356991-e832-4dfd-a6c5-e9b6d8e95980 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.766779] env[61868]: DEBUG oslo_vmware.rw_handles [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ddacd5-ae46-34f7-3c76-1835835a4316/disk-0.vmdk from lease info. {{(pid=61868) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 905.766779] env[61868]: DEBUG oslo_vmware.rw_handles [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ddacd5-ae46-34f7-3c76-1835835a4316/disk-0.vmdk for reading. {{(pid=61868) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 905.830435] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Releasing lock "refresh_cache-083140bf-233d-49ac-8247-cd93206f9200" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.830811] env[61868]: DEBUG nova.compute.manager [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Instance network_info: |[{"id": "79fe6cfd-0c91-4ec5-95b2-22aed62a0e0f", "address": "fa:16:3e:b7:c9:de", "network": {"id": "adc6ef1d-8fd6-4116-a2ff-34644d243811", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2090367503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f09bf553bf8b47e6846c59523d986edc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79fe6cfd-0c", "ovs_interfaceid": "79fe6cfd-0c91-4ec5-95b2-22aed62a0e0f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 905.834725] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b7:c9:de', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '79fe6cfd-0c91-4ec5-95b2-22aed62a0e0f', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 905.842887] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Creating folder: Project (f09bf553bf8b47e6846c59523d986edc). Parent ref: group-v281478. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 905.844508] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-72eb2fb5-8144-401f-983e-59b1968a4a06 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.859200] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Created folder: Project (f09bf553bf8b47e6846c59523d986edc) in parent group-v281478. [ 905.859416] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Creating folder: Instances. Parent ref: group-v281567. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 905.859665] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-85441682-5fba-4610-860a-28335ae0a2d8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.875141] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Created folder: Instances in parent group-v281567. [ 905.875141] env[61868]: DEBUG oslo.service.loopingcall [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 905.875141] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 905.875141] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c8b1c678-1388-4673-9a83-45804a616e6b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.896695] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 905.896695] env[61868]: value = "task-1315625" [ 905.896695] env[61868]: _type = "Task" [ 905.896695] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.906107] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315625, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.928866] env[61868]: DEBUG oslo_vmware.api [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315619, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.584032} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.929816] env[61868]: DEBUG nova.network.neutron [req-1de7ed7e-6052-4271-85fb-251feed25bd0 req-52f64ddb-8448-4899-a022-893ff7391b77 service nova] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Updated VIF entry in instance network info cache for port 29e8f7bd-db77-4071-b4c6-79e8338c2e82. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 905.930689] env[61868]: DEBUG nova.network.neutron [req-1de7ed7e-6052-4271-85fb-251feed25bd0 req-52f64ddb-8448-4899-a022-893ff7391b77 service nova] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Updating instance_info_cache with network_info: [{"id": "29e8f7bd-db77-4071-b4c6-79e8338c2e82", "address": "fa:16:3e:be:13:44", "network": {"id": "4befda35-5c88-4ac7-b8ef-128ddae97a7d", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-113207814-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "78a206c3daa14cc58e7fe8546f565ae0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd72ef32-a57c-43b0-93df-e8a030987d44", "external-id": "nsx-vlan-transportzone-340", "segmentation_id": 340, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap29e8f7bd-db", "ovs_interfaceid": "29e8f7bd-db77-4071-b4c6-79e8338c2e82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.933190] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 842c8606-632a-47d4-9a67-38d8e6ce6b4d/842c8606-632a-47d4-9a67-38d8e6ce6b4d.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 905.933190] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 905.933190] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1ab972e1-c1b4-4161-8b8d-2fe3e0bebc20 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.940383] env[61868]: DEBUG oslo_vmware.api [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 905.940383] env[61868]: value = "task-1315626" [ 905.940383] env[61868]: _type = "Task" [ 905.940383] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.954380] env[61868]: DEBUG oslo_concurrency.lockutils [None req-62918748-546f-4030-9046-6bb7ec988ab2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.900s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.954734] env[61868]: DEBUG oslo_vmware.api [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315626, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.956647] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a76b1c97-d8a9-4316-9001-c2415aa6d377 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.347s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.956963] env[61868]: DEBUG nova.objects.instance [None req-a76b1c97-d8a9-4316-9001-c2415aa6d377 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Lazy-loading 'resources' on Instance uuid c2693a43-4ea2-4ab0-8915-2fa544780e3c {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 905.992061] env[61868]: INFO nova.scheduler.client.report [None req-62918748-546f-4030-9046-6bb7ec988ab2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Deleted allocations for instance 52a5afc6-4e79-436c-bc94-b61ca9fb860c [ 905.993415] env[61868]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-570d285d-58fe-4f64-9f27-ad939484ff6f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.025938] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4913bdaa-6bad-444f-8690-03eeb16ef4b9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Lock "409cad1a-946d-4c58-aa57-1c0bf97fe63f" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.903s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.026650] env[61868]: INFO nova.compute.manager [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Took 33.94 seconds to build instance. [ 906.407834] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315625, 'name': CreateVM_Task, 'duration_secs': 0.455615} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.408019] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 906.408743] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.409960] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.409960] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 906.410236] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e8911d5-afd5-42ad-827a-542342e03a7d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.416163] env[61868]: DEBUG oslo_vmware.api [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Waiting for the task: (returnval){ [ 906.416163] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52edfab9-f40a-570b-b419-28c88d5b01b3" [ 906.416163] env[61868]: _type = "Task" [ 906.416163] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.427672] env[61868]: DEBUG oslo_vmware.api [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52edfab9-f40a-570b-b419-28c88d5b01b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.436485] env[61868]: DEBUG oslo_concurrency.lockutils [req-1de7ed7e-6052-4271-85fb-251feed25bd0 req-52f64ddb-8448-4899-a022-893ff7391b77 service nova] Releasing lock "refresh_cache-f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.436784] env[61868]: DEBUG nova.compute.manager [req-1de7ed7e-6052-4271-85fb-251feed25bd0 req-52f64ddb-8448-4899-a022-893ff7391b77 service nova] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Received event network-changed-29e8f7bd-db77-4071-b4c6-79e8338c2e82 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 906.436944] env[61868]: DEBUG nova.compute.manager [req-1de7ed7e-6052-4271-85fb-251feed25bd0 req-52f64ddb-8448-4899-a022-893ff7391b77 service nova] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Refreshing instance network info cache due to event network-changed-29e8f7bd-db77-4071-b4c6-79e8338c2e82. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 906.437186] env[61868]: DEBUG oslo_concurrency.lockutils [req-1de7ed7e-6052-4271-85fb-251feed25bd0 req-52f64ddb-8448-4899-a022-893ff7391b77 service nova] Acquiring lock "refresh_cache-f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.437333] env[61868]: DEBUG oslo_concurrency.lockutils [req-1de7ed7e-6052-4271-85fb-251feed25bd0 req-52f64ddb-8448-4899-a022-893ff7391b77 service nova] Acquired lock "refresh_cache-f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.437499] env[61868]: DEBUG nova.network.neutron [req-1de7ed7e-6052-4271-85fb-251feed25bd0 req-52f64ddb-8448-4899-a022-893ff7391b77 service nova] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Refreshing network info cache for port 29e8f7bd-db77-4071-b4c6-79e8338c2e82 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 906.453653] env[61868]: DEBUG oslo_vmware.api [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315626, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.0716} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.454127] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 906.455012] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d90d1182-b3ca-456d-acdd-06d7328c9bf6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.485051] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] 842c8606-632a-47d4-9a67-38d8e6ce6b4d/842c8606-632a-47d4-9a67-38d8e6ce6b4d.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 906.488433] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d09dd969-71d8-4bbc-8a9c-c8cd6ef68461 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.507293] env[61868]: DEBUG oslo_concurrency.lockutils [None req-62918748-546f-4030-9046-6bb7ec988ab2 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Lock "52a5afc6-4e79-436c-bc94-b61ca9fb860c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.307s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.512930] env[61868]: DEBUG oslo_vmware.api [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 906.512930] env[61868]: value = "task-1315627" [ 906.512930] env[61868]: _type = "Task" [ 906.512930] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.525959] env[61868]: DEBUG oslo_vmware.api [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315627, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.530553] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1fa116ea-c0eb-46d1-a014-147a4e4bd47a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Lock "8ea8e28b-3582-45f2-b4b5-84f624415a58" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.471s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.576505] env[61868]: DEBUG nova.compute.manager [req-3d694eb1-09af-45a6-b35c-ddad3dcd715f req-3dc54b84-e704-49f8-ad5d-468a7f7caba3 service nova] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Received event network-changed-79fe6cfd-0c91-4ec5-95b2-22aed62a0e0f {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 906.576711] env[61868]: DEBUG nova.compute.manager [req-3d694eb1-09af-45a6-b35c-ddad3dcd715f req-3dc54b84-e704-49f8-ad5d-468a7f7caba3 service nova] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Refreshing instance network info cache due to event network-changed-79fe6cfd-0c91-4ec5-95b2-22aed62a0e0f. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 906.576920] env[61868]: DEBUG oslo_concurrency.lockutils [req-3d694eb1-09af-45a6-b35c-ddad3dcd715f req-3dc54b84-e704-49f8-ad5d-468a7f7caba3 service nova] Acquiring lock "refresh_cache-083140bf-233d-49ac-8247-cd93206f9200" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.577090] env[61868]: DEBUG oslo_concurrency.lockutils [req-3d694eb1-09af-45a6-b35c-ddad3dcd715f req-3dc54b84-e704-49f8-ad5d-468a7f7caba3 service nova] Acquired lock "refresh_cache-083140bf-233d-49ac-8247-cd93206f9200" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.577550] env[61868]: DEBUG nova.network.neutron [req-3d694eb1-09af-45a6-b35c-ddad3dcd715f req-3dc54b84-e704-49f8-ad5d-468a7f7caba3 service nova] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Refreshing network info cache for port 79fe6cfd-0c91-4ec5-95b2-22aed62a0e0f {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 906.763029] env[61868]: DEBUG nova.network.neutron [-] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.797252] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfb33a9c-b620-4e17-95f1-038488b8067e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.806607] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28714516-ba65-4a4e-9c2b-fb3b1671b170 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.841832] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7622627-a4dd-4742-8dbf-35bb742c4646 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.851417] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2739237b-5dd4-4ba0-bd72-9e9ca9362aa3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.867693] env[61868]: DEBUG nova.compute.provider_tree [None req-a76b1c97-d8a9-4316-9001-c2415aa6d377 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 906.928098] env[61868]: DEBUG oslo_vmware.api [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52edfab9-f40a-570b-b419-28c88d5b01b3, 'name': SearchDatastore_Task, 'duration_secs': 0.020121} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.928453] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.928692] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 906.929113] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.929312] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.929574] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 906.929789] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-62897440-b5e9-4bfb-8a18-1973abae46df {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.943273] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 906.943508] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 906.944671] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c7e7c39-1264-4d19-8cbe-bb4b117be3b9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.952146] env[61868]: DEBUG oslo_vmware.api [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Waiting for the task: (returnval){ [ 906.952146] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52c173ed-37a3-576b-97ce-f7000b44acf1" [ 906.952146] env[61868]: _type = "Task" [ 906.952146] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.966913] env[61868]: DEBUG oslo_vmware.api [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52c173ed-37a3-576b-97ce-f7000b44acf1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.026149] env[61868]: DEBUG oslo_vmware.api [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315627, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.210196] env[61868]: DEBUG nova.network.neutron [req-1de7ed7e-6052-4271-85fb-251feed25bd0 req-52f64ddb-8448-4899-a022-893ff7391b77 service nova] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Updated VIF entry in instance network info cache for port 29e8f7bd-db77-4071-b4c6-79e8338c2e82. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 907.210196] env[61868]: DEBUG nova.network.neutron [req-1de7ed7e-6052-4271-85fb-251feed25bd0 req-52f64ddb-8448-4899-a022-893ff7391b77 service nova] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Updating instance_info_cache with network_info: [{"id": "29e8f7bd-db77-4071-b4c6-79e8338c2e82", "address": "fa:16:3e:be:13:44", "network": {"id": "4befda35-5c88-4ac7-b8ef-128ddae97a7d", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-113207814-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.143", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "78a206c3daa14cc58e7fe8546f565ae0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd72ef32-a57c-43b0-93df-e8a030987d44", "external-id": "nsx-vlan-transportzone-340", "segmentation_id": 340, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap29e8f7bd-db", "ovs_interfaceid": "29e8f7bd-db77-4071-b4c6-79e8338c2e82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.263167] env[61868]: DEBUG nova.compute.manager [req-75caa470-3f76-4dff-8b03-24320cd421cd req-92565418-8603-4f2b-96b5-58fd182b7ef8 service nova] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Received event network-changed-29e8f7bd-db77-4071-b4c6-79e8338c2e82 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 907.264253] env[61868]: DEBUG nova.compute.manager [req-75caa470-3f76-4dff-8b03-24320cd421cd req-92565418-8603-4f2b-96b5-58fd182b7ef8 service nova] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Refreshing instance network info cache due to event network-changed-29e8f7bd-db77-4071-b4c6-79e8338c2e82. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 907.264253] env[61868]: DEBUG oslo_concurrency.lockutils [req-75caa470-3f76-4dff-8b03-24320cd421cd req-92565418-8603-4f2b-96b5-58fd182b7ef8 service nova] Acquiring lock "refresh_cache-f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.268647] env[61868]: INFO nova.compute.manager [-] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Took 1.64 seconds to deallocate network for instance. [ 907.372565] env[61868]: DEBUG nova.scheduler.client.report [None req-a76b1c97-d8a9-4316-9001-c2415aa6d377 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 907.464762] env[61868]: DEBUG oslo_vmware.api [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52c173ed-37a3-576b-97ce-f7000b44acf1, 'name': SearchDatastore_Task, 'duration_secs': 0.024971} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.465680] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10ca9086-5733-40a6-af3e-6780a002e1b9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.472550] env[61868]: DEBUG oslo_vmware.api [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Waiting for the task: (returnval){ [ 907.472550] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52d8edb4-8425-2bea-2226-c0704b5f3772" [ 907.472550] env[61868]: _type = "Task" [ 907.472550] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.485969] env[61868]: DEBUG oslo_vmware.api [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52d8edb4-8425-2bea-2226-c0704b5f3772, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.525435] env[61868]: DEBUG oslo_vmware.api [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315627, 'name': ReconfigVM_Task, 'duration_secs': 0.662641} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.525742] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Reconfigured VM instance instance-0000004e to attach disk [datastore2] 842c8606-632a-47d4-9a67-38d8e6ce6b4d/842c8606-632a-47d4-9a67-38d8e6ce6b4d.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 907.526446] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1251b00e-57a2-4063-a4dc-c8b6e328de3d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.534084] env[61868]: DEBUG oslo_vmware.api [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 907.534084] env[61868]: value = "task-1315628" [ 907.534084] env[61868]: _type = "Task" [ 907.534084] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.544565] env[61868]: DEBUG oslo_vmware.api [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315628, 'name': Rename_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.548966] env[61868]: DEBUG nova.network.neutron [req-3d694eb1-09af-45a6-b35c-ddad3dcd715f req-3dc54b84-e704-49f8-ad5d-468a7f7caba3 service nova] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Updated VIF entry in instance network info cache for port 79fe6cfd-0c91-4ec5-95b2-22aed62a0e0f. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 907.548966] env[61868]: DEBUG nova.network.neutron [req-3d694eb1-09af-45a6-b35c-ddad3dcd715f req-3dc54b84-e704-49f8-ad5d-468a7f7caba3 service nova] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Updating instance_info_cache with network_info: [{"id": "79fe6cfd-0c91-4ec5-95b2-22aed62a0e0f", "address": "fa:16:3e:b7:c9:de", "network": {"id": "adc6ef1d-8fd6-4116-a2ff-34644d243811", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2090367503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f09bf553bf8b47e6846c59523d986edc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79fe6cfd-0c", "ovs_interfaceid": "79fe6cfd-0c91-4ec5-95b2-22aed62a0e0f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.713018] env[61868]: DEBUG oslo_concurrency.lockutils [req-1de7ed7e-6052-4271-85fb-251feed25bd0 req-52f64ddb-8448-4899-a022-893ff7391b77 service nova] Releasing lock "refresh_cache-f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.713018] env[61868]: DEBUG oslo_concurrency.lockutils [req-75caa470-3f76-4dff-8b03-24320cd421cd req-92565418-8603-4f2b-96b5-58fd182b7ef8 service nova] Acquired lock "refresh_cache-f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.713018] env[61868]: DEBUG nova.network.neutron [req-75caa470-3f76-4dff-8b03-24320cd421cd req-92565418-8603-4f2b-96b5-58fd182b7ef8 service nova] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Refreshing network info cache for port 29e8f7bd-db77-4071-b4c6-79e8338c2e82 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 907.776880] env[61868]: DEBUG oslo_concurrency.lockutils [None req-780bfa57-3788-4c6b-93f4-f83d47616fd0 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.780115] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Acquiring lock "6ce39c07-939c-49f3-8871-6e2163709f4a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.782545] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Lock "6ce39c07-939c-49f3-8871-6e2163709f4a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.854264] env[61868]: INFO nova.compute.manager [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Rescuing [ 907.855493] env[61868]: DEBUG oslo_concurrency.lockutils [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Acquiring lock "refresh_cache-8ea8e28b-3582-45f2-b4b5-84f624415a58" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.855493] env[61868]: DEBUG oslo_concurrency.lockutils [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Acquired lock "refresh_cache-8ea8e28b-3582-45f2-b4b5-84f624415a58" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.855493] env[61868]: DEBUG nova.network.neutron [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 907.877665] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a76b1c97-d8a9-4316-9001-c2415aa6d377 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.921s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.882012] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.338s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.887108] env[61868]: INFO nova.compute.claims [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 907.909543] env[61868]: INFO nova.scheduler.client.report [None req-a76b1c97-d8a9-4316-9001-c2415aa6d377 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Deleted allocations for instance c2693a43-4ea2-4ab0-8915-2fa544780e3c [ 907.988578] env[61868]: DEBUG oslo_vmware.api [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52d8edb4-8425-2bea-2226-c0704b5f3772, 'name': SearchDatastore_Task, 'duration_secs': 0.013051} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.988578] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.988790] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 083140bf-233d-49ac-8247-cd93206f9200/083140bf-233d-49ac-8247-cd93206f9200.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 907.989808] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-65e7387a-3182-42bc-b25a-12b244f49edf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.999620] env[61868]: DEBUG oslo_vmware.api [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Waiting for the task: (returnval){ [ 907.999620] env[61868]: value = "task-1315629" [ 907.999620] env[61868]: _type = "Task" [ 907.999620] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.013901] env[61868]: DEBUG oslo_vmware.api [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Task: {'id': task-1315629, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.046122] env[61868]: DEBUG oslo_vmware.api [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315628, 'name': Rename_Task, 'duration_secs': 0.190199} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.046473] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 908.046740] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-17ea4788-477c-407c-bd61-5030bdd33a08 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.053481] env[61868]: DEBUG oslo_concurrency.lockutils [req-3d694eb1-09af-45a6-b35c-ddad3dcd715f req-3dc54b84-e704-49f8-ad5d-468a7f7caba3 service nova] Releasing lock "refresh_cache-083140bf-233d-49ac-8247-cd93206f9200" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.056460] env[61868]: DEBUG oslo_vmware.api [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 908.056460] env[61868]: value = "task-1315630" [ 908.056460] env[61868]: _type = "Task" [ 908.056460] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.070640] env[61868]: DEBUG oslo_vmware.api [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315630, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.283791] env[61868]: DEBUG nova.compute.manager [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 908.418564] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a76b1c97-d8a9-4316-9001-c2415aa6d377 tempest-ListServersNegativeTestJSON-2126701456 tempest-ListServersNegativeTestJSON-2126701456-project-member] Lock "c2693a43-4ea2-4ab0-8915-2fa544780e3c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.944s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.515906] env[61868]: DEBUG oslo_vmware.api [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Task: {'id': task-1315629, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.568265] env[61868]: DEBUG oslo_concurrency.lockutils [None req-001b27d5-1800-43bb-b6d4-35be0ea48f57 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Acquiring lock "409cad1a-946d-4c58-aa57-1c0bf97fe63f" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.568642] env[61868]: DEBUG oslo_concurrency.lockutils [None req-001b27d5-1800-43bb-b6d4-35be0ea48f57 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Lock "409cad1a-946d-4c58-aa57-1c0bf97fe63f" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.577352] env[61868]: DEBUG oslo_vmware.api [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315630, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.650624] env[61868]: DEBUG nova.network.neutron [req-75caa470-3f76-4dff-8b03-24320cd421cd req-92565418-8603-4f2b-96b5-58fd182b7ef8 service nova] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Updated VIF entry in instance network info cache for port 29e8f7bd-db77-4071-b4c6-79e8338c2e82. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 908.651031] env[61868]: DEBUG nova.network.neutron [req-75caa470-3f76-4dff-8b03-24320cd421cd req-92565418-8603-4f2b-96b5-58fd182b7ef8 service nova] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Updating instance_info_cache with network_info: [{"id": "29e8f7bd-db77-4071-b4c6-79e8338c2e82", "address": "fa:16:3e:be:13:44", "network": {"id": "4befda35-5c88-4ac7-b8ef-128ddae97a7d", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-113207814-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "78a206c3daa14cc58e7fe8546f565ae0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd72ef32-a57c-43b0-93df-e8a030987d44", "external-id": "nsx-vlan-transportzone-340", "segmentation_id": 340, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap29e8f7bd-db", "ovs_interfaceid": "29e8f7bd-db77-4071-b4c6-79e8338c2e82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.719294] env[61868]: DEBUG nova.network.neutron [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Updating instance_info_cache with network_info: [{"id": "c27d3e06-cf3c-4ef8-ba0b-e6f2ca824646", "address": "fa:16:3e:3e:8f:76", "network": {"id": "094eaf15-c932-4d4f-8736-fe5239160507", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1905020504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "7541757d43c74b93acf728aa2fb0f425", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc27d3e06-cf", "ovs_interfaceid": "c27d3e06-cf3c-4ef8-ba0b-e6f2ca824646", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.806360] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.869642] env[61868]: DEBUG nova.compute.manager [req-d866c2d9-aa07-46ca-9ee8-9ae8f0c6f69b req-cd8be08b-5ba6-44e1-b20a-daea7e3e8d63 service nova] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Received event network-vif-deleted-e76c5ee1-88d7-4366-a731-7b793639fcef {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 909.015030] env[61868]: DEBUG oslo_vmware.api [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Task: {'id': task-1315629, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.687269} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.015030] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 083140bf-233d-49ac-8247-cd93206f9200/083140bf-233d-49ac-8247-cd93206f9200.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 909.015030] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 909.015030] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-eec2cd9c-655e-43f4-aa5b-5166325530f5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.026587] env[61868]: DEBUG oslo_vmware.api [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Waiting for the task: (returnval){ [ 909.026587] env[61868]: value = "task-1315631" [ 909.026587] env[61868]: _type = "Task" [ 909.026587] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.035771] env[61868]: DEBUG oslo_vmware.api [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Task: {'id': task-1315631, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.070588] env[61868]: DEBUG oslo_vmware.api [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315630, 'name': PowerOnVM_Task, 'duration_secs': 0.605922} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.070588] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 909.070588] env[61868]: INFO nova.compute.manager [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Took 8.91 seconds to spawn the instance on the hypervisor. [ 909.070588] env[61868]: DEBUG nova.compute.manager [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 909.070588] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90f0a5b6-314e-4aa0-bf01-740df89312e2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.073107] env[61868]: DEBUG nova.compute.utils [None req-001b27d5-1800-43bb-b6d4-35be0ea48f57 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 909.156077] env[61868]: DEBUG oslo_concurrency.lockutils [req-75caa470-3f76-4dff-8b03-24320cd421cd req-92565418-8603-4f2b-96b5-58fd182b7ef8 service nova] Releasing lock "refresh_cache-f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.191539] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-657d9edb-0ad5-4c82-8632-a53dad0c1e2f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.200026] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-507541b3-7a2d-46c7-88a4-ce20fa0f5fb1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.237514] env[61868]: DEBUG oslo_concurrency.lockutils [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Releasing lock "refresh_cache-8ea8e28b-3582-45f2-b4b5-84f624415a58" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.240675] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22f47d67-035c-4df3-8fa8-b67380c16449 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.248887] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-703cd271-c7eb-4cce-97eb-bdcb638335a5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.266287] env[61868]: DEBUG nova.compute.provider_tree [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 909.533364] env[61868]: DEBUG oslo_vmware.api [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Task: {'id': task-1315631, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072862} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.533651] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 909.534508] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aafbab8b-f905-451f-829c-c006a3af8789 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.558417] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Reconfiguring VM instance instance-00000035 to attach disk [datastore1] 083140bf-233d-49ac-8247-cd93206f9200/083140bf-233d-49ac-8247-cd93206f9200.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 909.559546] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2b7fdcd2-8286-4e08-ad1b-ae514200e43e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.576391] env[61868]: DEBUG oslo_concurrency.lockutils [None req-001b27d5-1800-43bb-b6d4-35be0ea48f57 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Lock "409cad1a-946d-4c58-aa57-1c0bf97fe63f" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.587769] env[61868]: DEBUG oslo_vmware.api [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Waiting for the task: (returnval){ [ 909.587769] env[61868]: value = "task-1315632" [ 909.587769] env[61868]: _type = "Task" [ 909.587769] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.592759] env[61868]: INFO nova.compute.manager [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Took 36.24 seconds to build instance. [ 909.603403] env[61868]: DEBUG oslo_vmware.api [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Task: {'id': task-1315632, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.736144] env[61868]: DEBUG nova.compute.manager [req-97ce70b3-3226-4a96-8f86-0ff114903eae req-625f77f9-0d0a-49a1-9419-c789adc9abe6 service nova] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Received event network-changed-29e8f7bd-db77-4071-b4c6-79e8338c2e82 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 909.736344] env[61868]: DEBUG nova.compute.manager [req-97ce70b3-3226-4a96-8f86-0ff114903eae req-625f77f9-0d0a-49a1-9419-c789adc9abe6 service nova] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Refreshing instance network info cache due to event network-changed-29e8f7bd-db77-4071-b4c6-79e8338c2e82. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 909.736560] env[61868]: DEBUG oslo_concurrency.lockutils [req-97ce70b3-3226-4a96-8f86-0ff114903eae req-625f77f9-0d0a-49a1-9419-c789adc9abe6 service nova] Acquiring lock "refresh_cache-f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.736734] env[61868]: DEBUG oslo_concurrency.lockutils [req-97ce70b3-3226-4a96-8f86-0ff114903eae req-625f77f9-0d0a-49a1-9419-c789adc9abe6 service nova] Acquired lock "refresh_cache-f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.736912] env[61868]: DEBUG nova.network.neutron [req-97ce70b3-3226-4a96-8f86-0ff114903eae req-625f77f9-0d0a-49a1-9419-c789adc9abe6 service nova] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Refreshing network info cache for port 29e8f7bd-db77-4071-b4c6-79e8338c2e82 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 909.770985] env[61868]: DEBUG nova.scheduler.client.report [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 909.780567] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 909.780696] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2c1dd818-5658-4ec3-9e0c-9a0ee9a3cff9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.789788] env[61868]: DEBUG oslo_vmware.api [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 909.789788] env[61868]: value = "task-1315633" [ 909.789788] env[61868]: _type = "Task" [ 909.789788] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.799523] env[61868]: DEBUG oslo_vmware.api [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315633, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.922165] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4752328e-63ce-4df8-ad78-8a6198a20895 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Acquiring lock "f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.922484] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4752328e-63ce-4df8-ad78-8a6198a20895 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Lock "f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.922666] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4752328e-63ce-4df8-ad78-8a6198a20895 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Acquiring lock "f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.922859] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4752328e-63ce-4df8-ad78-8a6198a20895 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Lock "f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.923046] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4752328e-63ce-4df8-ad78-8a6198a20895 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Lock "f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.926072] env[61868]: INFO nova.compute.manager [None req-4752328e-63ce-4df8-ad78-8a6198a20895 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Terminating instance [ 909.929011] env[61868]: DEBUG nova.compute.manager [None req-4752328e-63ce-4df8-ad78-8a6198a20895 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 909.929273] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-4752328e-63ce-4df8-ad78-8a6198a20895 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 909.930397] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8dbf06b-7423-4690-ac20-c0ee319160cd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.941104] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-4752328e-63ce-4df8-ad78-8a6198a20895 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 909.941270] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b6f48f15-7d79-43af-899e-c148c9ebb2e0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.953829] env[61868]: DEBUG oslo_vmware.api [None req-4752328e-63ce-4df8-ad78-8a6198a20895 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Waiting for the task: (returnval){ [ 909.953829] env[61868]: value = "task-1315634" [ 909.953829] env[61868]: _type = "Task" [ 909.953829] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.964920] env[61868]: DEBUG oslo_vmware.api [None req-4752328e-63ce-4df8-ad78-8a6198a20895 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': task-1315634, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.095750] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f491b12c-033d-4586-88a1-9bf7a9b99093 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "842c8606-632a-47d4-9a67-38d8e6ce6b4d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.749s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.102127] env[61868]: DEBUG oslo_vmware.api [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Task: {'id': task-1315632, 'name': ReconfigVM_Task, 'duration_secs': 0.42447} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.102432] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Reconfigured VM instance instance-00000035 to attach disk [datastore1] 083140bf-233d-49ac-8247-cd93206f9200/083140bf-233d-49ac-8247-cd93206f9200.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 910.103119] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-69fbf544-7837-4db6-bd3f-7c82a55cabf8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.111619] env[61868]: DEBUG oslo_vmware.api [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Waiting for the task: (returnval){ [ 910.111619] env[61868]: value = "task-1315635" [ 910.111619] env[61868]: _type = "Task" [ 910.111619] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.122575] env[61868]: DEBUG oslo_vmware.api [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Task: {'id': task-1315635, 'name': Rename_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.276971] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.395s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.277202] env[61868]: DEBUG nova.compute.manager [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 910.280175] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.164s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.282102] env[61868]: INFO nova.compute.claims [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 910.302787] env[61868]: DEBUG oslo_vmware.api [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315633, 'name': PowerOffVM_Task, 'duration_secs': 0.282443} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.303746] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 910.304647] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad3d541f-5487-4bf8-bdaa-dc0ce62d621f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.331370] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cab0ee9-909b-4d11-b29b-7c10e809572b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.378550] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 910.379961] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c4bb9252-aa39-42d9-9a4f-2102e6f7e5bd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.387209] env[61868]: DEBUG oslo_vmware.api [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 910.387209] env[61868]: value = "task-1315636" [ 910.387209] env[61868]: _type = "Task" [ 910.387209] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.398823] env[61868]: DEBUG oslo_vmware.api [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315636, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.466604] env[61868]: DEBUG oslo_vmware.api [None req-4752328e-63ce-4df8-ad78-8a6198a20895 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': task-1315634, 'name': PowerOffVM_Task, 'duration_secs': 0.387675} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.469401] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-4752328e-63ce-4df8-ad78-8a6198a20895 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 910.469606] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-4752328e-63ce-4df8-ad78-8a6198a20895 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 910.469906] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f118e67a-2f9b-4fc7-8f76-5195d37b4040 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.511490] env[61868]: DEBUG nova.network.neutron [req-97ce70b3-3226-4a96-8f86-0ff114903eae req-625f77f9-0d0a-49a1-9419-c789adc9abe6 service nova] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Updated VIF entry in instance network info cache for port 29e8f7bd-db77-4071-b4c6-79e8338c2e82. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 910.511490] env[61868]: DEBUG nova.network.neutron [req-97ce70b3-3226-4a96-8f86-0ff114903eae req-625f77f9-0d0a-49a1-9419-c789adc9abe6 service nova] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Updating instance_info_cache with network_info: [{"id": "29e8f7bd-db77-4071-b4c6-79e8338c2e82", "address": "fa:16:3e:be:13:44", "network": {"id": "4befda35-5c88-4ac7-b8ef-128ddae97a7d", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-113207814-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "78a206c3daa14cc58e7fe8546f565ae0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd72ef32-a57c-43b0-93df-e8a030987d44", "external-id": "nsx-vlan-transportzone-340", "segmentation_id": 340, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap29e8f7bd-db", "ovs_interfaceid": "29e8f7bd-db77-4071-b4c6-79e8338c2e82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.547432] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-4752328e-63ce-4df8-ad78-8a6198a20895 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 910.547748] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-4752328e-63ce-4df8-ad78-8a6198a20895 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 910.548010] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-4752328e-63ce-4df8-ad78-8a6198a20895 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Deleting the datastore file [datastore1] f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 910.548319] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fd172029-0191-493f-ae3c-7a2260964d83 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.556945] env[61868]: DEBUG oslo_vmware.api [None req-4752328e-63ce-4df8-ad78-8a6198a20895 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Waiting for the task: (returnval){ [ 910.556945] env[61868]: value = "task-1315638" [ 910.556945] env[61868]: _type = "Task" [ 910.556945] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.566248] env[61868]: DEBUG oslo_vmware.api [None req-4752328e-63ce-4df8-ad78-8a6198a20895 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': task-1315638, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.622782] env[61868]: DEBUG oslo_vmware.api [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Task: {'id': task-1315635, 'name': Rename_Task, 'duration_secs': 0.247254} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.622782] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 910.623049] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7d5caed4-657c-4be3-bbf1-23922631daf6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.631210] env[61868]: DEBUG oslo_vmware.api [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Waiting for the task: (returnval){ [ 910.631210] env[61868]: value = "task-1315639" [ 910.631210] env[61868]: _type = "Task" [ 910.631210] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.640679] env[61868]: DEBUG oslo_vmware.api [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Task: {'id': task-1315639, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.664145] env[61868]: DEBUG oslo_concurrency.lockutils [None req-001b27d5-1800-43bb-b6d4-35be0ea48f57 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Acquiring lock "409cad1a-946d-4c58-aa57-1c0bf97fe63f" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.664452] env[61868]: DEBUG oslo_concurrency.lockutils [None req-001b27d5-1800-43bb-b6d4-35be0ea48f57 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Lock "409cad1a-946d-4c58-aa57-1c0bf97fe63f" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.664696] env[61868]: INFO nova.compute.manager [None req-001b27d5-1800-43bb-b6d4-35be0ea48f57 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Attaching volume fdcd4a1e-341a-4607-b405-3e3c2a834712 to /dev/sdc [ 910.718020] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c501edd4-bfef-46e2-8271-a46855fa48cc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.727127] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20173d55-4e8e-45ad-9ba2-835f827918c0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.745975] env[61868]: DEBUG nova.virt.block_device [None req-001b27d5-1800-43bb-b6d4-35be0ea48f57 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Updating existing volume attachment record: 1eda38a6-925b-4d8e-9e0b-11e5f10eb94d {{(pid=61868) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 910.789022] env[61868]: DEBUG nova.compute.utils [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 910.791940] env[61868]: DEBUG nova.compute.manager [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 910.794021] env[61868]: DEBUG nova.network.neutron [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 910.843376] env[61868]: DEBUG nova.policy [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0fa71d6f8941441d89709e7e3af1ed9f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c8440ce785c344d793dd513b5f008d65', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 910.899397] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] VM already powered off {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 910.899744] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 910.900150] env[61868]: DEBUG oslo_concurrency.lockutils [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.900385] env[61868]: DEBUG oslo_concurrency.lockutils [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.900653] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 910.900991] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c627c699-e3ec-4288-893b-8f1b7e1621b3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.914557] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 910.914786] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 910.915778] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-947285b5-9c05-455d-a999-44e8828881bf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.925333] env[61868]: DEBUG oslo_vmware.api [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 910.925333] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]521cddfc-6341-6566-986e-8039250a17dd" [ 910.925333] env[61868]: _type = "Task" [ 910.925333] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.931957] env[61868]: DEBUG oslo_vmware.api [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]521cddfc-6341-6566-986e-8039250a17dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.014888] env[61868]: DEBUG oslo_concurrency.lockutils [req-97ce70b3-3226-4a96-8f86-0ff114903eae req-625f77f9-0d0a-49a1-9419-c789adc9abe6 service nova] Releasing lock "refresh_cache-f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.071385] env[61868]: DEBUG oslo_vmware.api [None req-4752328e-63ce-4df8-ad78-8a6198a20895 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Task: {'id': task-1315638, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.324862} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.071737] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-4752328e-63ce-4df8-ad78-8a6198a20895 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 911.071982] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-4752328e-63ce-4df8-ad78-8a6198a20895 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 911.072200] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-4752328e-63ce-4df8-ad78-8a6198a20895 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 911.072429] env[61868]: INFO nova.compute.manager [None req-4752328e-63ce-4df8-ad78-8a6198a20895 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Took 1.14 seconds to destroy the instance on the hypervisor. [ 911.072749] env[61868]: DEBUG oslo.service.loopingcall [None req-4752328e-63ce-4df8-ad78-8a6198a20895 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 911.073422] env[61868]: DEBUG nova.compute.manager [-] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 911.073523] env[61868]: DEBUG nova.network.neutron [-] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 911.146284] env[61868]: DEBUG oslo_vmware.api [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Task: {'id': task-1315639, 'name': PowerOnVM_Task, 'duration_secs': 0.511495} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.150025] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 911.150025] env[61868]: INFO nova.compute.manager [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Took 99.46 seconds to spawn the instance on the hypervisor. [ 911.150025] env[61868]: DEBUG nova.compute.manager [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 911.150025] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f7846cf-d33d-4c74-9405-e17d21ed6e6c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.295022] env[61868]: DEBUG nova.compute.manager [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 911.336505] env[61868]: DEBUG nova.network.neutron [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Successfully created port: cb19d64d-19c9-4d48-8d70-00b3440d9127 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 911.440645] env[61868]: DEBUG oslo_vmware.api [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]521cddfc-6341-6566-986e-8039250a17dd, 'name': SearchDatastore_Task, 'duration_secs': 0.012853} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.441834] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-828713a0-1599-4b4c-af43-2c98f193ec35 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.449829] env[61868]: DEBUG oslo_vmware.api [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 911.449829] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52b4905e-ac45-d277-b555-afdf0b33cb30" [ 911.449829] env[61868]: _type = "Task" [ 911.449829] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.480065] env[61868]: DEBUG oslo_vmware.api [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52b4905e-ac45-d277-b555-afdf0b33cb30, 'name': SearchDatastore_Task, 'duration_secs': 0.013752} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.480961] env[61868]: DEBUG oslo_concurrency.lockutils [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.481368] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 8ea8e28b-3582-45f2-b4b5-84f624415a58/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d-rescue.vmdk. {{(pid=61868) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 911.481804] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2b178529-90bd-4848-ac7a-769dc9916fad {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.493881] env[61868]: DEBUG oslo_vmware.api [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 911.493881] env[61868]: value = "task-1315641" [ 911.493881] env[61868]: _type = "Task" [ 911.493881] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.509700] env[61868]: DEBUG oslo_vmware.api [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315641, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.663670] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-212b41f0-9a60-44d4-bced-d8871975db37 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.666456] env[61868]: DEBUG nova.compute.utils [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Conflict updating instance 083140bf-233d-49ac-8247-cd93206f9200. Expected: {'task_state': ['spawning']}. Actual: {'task_state': 'deleting'} {{(pid=61868) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 911.668216] env[61868]: DEBUG nova.compute.manager [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Instance disappeared during build. {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2513}} [ 911.668481] env[61868]: DEBUG nova.compute.manager [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Unplugging VIFs for instance {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 911.668683] env[61868]: DEBUG nova.compute.manager [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61868) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 911.668868] env[61868]: DEBUG nova.compute.manager [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 911.669066] env[61868]: DEBUG nova.network.neutron [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 911.673992] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-514db1a2-c83b-4524-b410-85c30952230d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.706952] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29583158-4d8b-4154-a288-f225bfc9c2da {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.715863] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-058892ee-040a-4a55-940a-637d601d5ff1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.733847] env[61868]: DEBUG nova.compute.provider_tree [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 912.009418] env[61868]: DEBUG oslo_vmware.api [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315641, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.129855] env[61868]: DEBUG nova.network.neutron [-] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.229930] env[61868]: DEBUG nova.compute.manager [req-2f2fe875-812b-4e69-a094-bd22377f1780 req-17782f62-defd-4875-8931-37654d384500 service nova] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Received event network-vif-deleted-29e8f7bd-db77-4071-b4c6-79e8338c2e82 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 912.237717] env[61868]: DEBUG nova.scheduler.client.report [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 912.308302] env[61868]: DEBUG nova.compute.manager [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 912.335900] env[61868]: DEBUG nova.virt.hardware [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 912.336168] env[61868]: DEBUG nova.virt.hardware [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 912.336332] env[61868]: DEBUG nova.virt.hardware [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 912.336522] env[61868]: DEBUG nova.virt.hardware [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 912.337299] env[61868]: DEBUG nova.virt.hardware [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 912.337685] env[61868]: DEBUG nova.virt.hardware [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 912.338384] env[61868]: DEBUG nova.virt.hardware [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 912.338804] env[61868]: DEBUG nova.virt.hardware [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 912.339118] env[61868]: DEBUG nova.virt.hardware [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 912.339654] env[61868]: DEBUG nova.virt.hardware [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 912.340118] env[61868]: DEBUG nova.virt.hardware [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 912.341446] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a94eb622-75b3-44bc-bdd3-78f25d0b49d8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.353336] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dab1c3b9-c762-4958-8cbf-b15ae75f756e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.381648] env[61868]: DEBUG nova.compute.manager [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Stashing vm_state: active {{(pid=61868) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 912.513443] env[61868]: DEBUG oslo_vmware.api [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315641, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.669867} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.513754] env[61868]: INFO nova.virt.vmwareapi.ds_util [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 8ea8e28b-3582-45f2-b4b5-84f624415a58/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d-rescue.vmdk. [ 912.515133] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65f5b702-2864-49f7-8178-ac916fe175de {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.565369] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] 8ea8e28b-3582-45f2-b4b5-84f624415a58/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d-rescue.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 912.566076] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-872d2e8b-9331-4a04-ad5d-1c6b55a9e5da {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.588984] env[61868]: DEBUG oslo_vmware.api [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 912.588984] env[61868]: value = "task-1315642" [ 912.588984] env[61868]: _type = "Task" [ 912.588984] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.599921] env[61868]: DEBUG oslo_vmware.api [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315642, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.632704] env[61868]: INFO nova.compute.manager [-] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Took 1.56 seconds to deallocate network for instance. [ 912.746209] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.464s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.746209] env[61868]: DEBUG nova.compute.manager [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 912.748172] env[61868]: DEBUG oslo_concurrency.lockutils [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.133s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.749995] env[61868]: INFO nova.compute.claims [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 912.916259] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.930527] env[61868]: DEBUG nova.compute.manager [req-0012e2dc-ec35-4255-af97-7a4634f7c3fb req-eca38b5e-4f22-45a3-a25e-02be9c5771d8 service nova] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Received event network-vif-plugged-cb19d64d-19c9-4d48-8d70-00b3440d9127 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 912.931779] env[61868]: DEBUG oslo_concurrency.lockutils [req-0012e2dc-ec35-4255-af97-7a4634f7c3fb req-eca38b5e-4f22-45a3-a25e-02be9c5771d8 service nova] Acquiring lock "e59a182d-97b2-454a-bc40-8afb0839324f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.931779] env[61868]: DEBUG oslo_concurrency.lockutils [req-0012e2dc-ec35-4255-af97-7a4634f7c3fb req-eca38b5e-4f22-45a3-a25e-02be9c5771d8 service nova] Lock "e59a182d-97b2-454a-bc40-8afb0839324f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.933684] env[61868]: DEBUG oslo_concurrency.lockutils [req-0012e2dc-ec35-4255-af97-7a4634f7c3fb req-eca38b5e-4f22-45a3-a25e-02be9c5771d8 service nova] Lock "e59a182d-97b2-454a-bc40-8afb0839324f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.002s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.934042] env[61868]: DEBUG nova.compute.manager [req-0012e2dc-ec35-4255-af97-7a4634f7c3fb req-eca38b5e-4f22-45a3-a25e-02be9c5771d8 service nova] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] No waiting events found dispatching network-vif-plugged-cb19d64d-19c9-4d48-8d70-00b3440d9127 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 912.934349] env[61868]: WARNING nova.compute.manager [req-0012e2dc-ec35-4255-af97-7a4634f7c3fb req-eca38b5e-4f22-45a3-a25e-02be9c5771d8 service nova] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Received unexpected event network-vif-plugged-cb19d64d-19c9-4d48-8d70-00b3440d9127 for instance with vm_state building and task_state spawning. [ 913.029254] env[61868]: DEBUG nova.network.neutron [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.049290] env[61868]: DEBUG nova.network.neutron [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Successfully updated port: cb19d64d-19c9-4d48-8d70-00b3440d9127 {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 913.102704] env[61868]: DEBUG oslo_vmware.api [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315642, 'name': ReconfigVM_Task, 'duration_secs': 0.483203} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.103144] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Reconfigured VM instance instance-0000004d to attach disk [datastore1] 8ea8e28b-3582-45f2-b4b5-84f624415a58/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d-rescue.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 913.104179] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fc1414f-2115-4155-8832-250b08fa9c22 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.138466] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d7f6eb94-22dc-4815-99a3-1403aada3e25 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.149473] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4752328e-63ce-4df8-ad78-8a6198a20895 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.158155] env[61868]: DEBUG oslo_vmware.api [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 913.158155] env[61868]: value = "task-1315644" [ 913.158155] env[61868]: _type = "Task" [ 913.158155] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.169647] env[61868]: DEBUG oslo_vmware.api [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315644, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.254747] env[61868]: DEBUG nova.compute.utils [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 913.258441] env[61868]: DEBUG nova.compute.manager [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Not allocating networking since 'none' was specified. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 913.533247] env[61868]: INFO nova.compute.manager [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Took 1.86 seconds to deallocate network for instance. [ 913.553408] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "refresh_cache-e59a182d-97b2-454a-bc40-8afb0839324f" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.553573] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquired lock "refresh_cache-e59a182d-97b2-454a-bc40-8afb0839324f" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.553792] env[61868]: DEBUG nova.network.neutron [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 913.668411] env[61868]: DEBUG oslo_vmware.api [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315644, 'name': ReconfigVM_Task, 'duration_secs': 0.260537} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.668701] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 913.668974] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1b21b71e-8d6f-4bb1-9548-e989b59d76d5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.677350] env[61868]: DEBUG oslo_vmware.api [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 913.677350] env[61868]: value = "task-1315645" [ 913.677350] env[61868]: _type = "Task" [ 913.677350] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.687213] env[61868]: DEBUG oslo_vmware.api [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315645, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.758993] env[61868]: DEBUG nova.compute.manager [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 914.092790] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22c1edd1-839f-46c7-98cd-b32ce25f1b6e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.096665] env[61868]: DEBUG nova.network.neutron [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 914.106902] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89ce0f63-c235-40bc-bc99-fdda9e51bf25 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.146519] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df446799-35e1-45a7-b438-8cb017189304 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.158281] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b7d94b9-7a77-4c9c-8bf3-bb89ddd2864d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.175421] env[61868]: DEBUG nova.compute.provider_tree [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 914.189739] env[61868]: DEBUG oslo_vmware.api [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315645, 'name': PowerOnVM_Task, 'duration_secs': 0.505255} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.190065] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 914.193248] env[61868]: DEBUG nova.compute.manager [None req-cf843a84-e08a-4581-b893-85a98c92c66a tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 914.193457] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65e077c2-66d5-4955-92bd-64b3bef86f14 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.287863] env[61868]: DEBUG nova.network.neutron [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Updating instance_info_cache with network_info: [{"id": "cb19d64d-19c9-4d48-8d70-00b3440d9127", "address": "fa:16:3e:d4:cc:5c", "network": {"id": "3535e048-fe57-4cbc-9489-0b77484069bd", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-748518849-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c8440ce785c344d793dd513b5f008d65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb19d64d-19", "ovs_interfaceid": "cb19d64d-19c9-4d48-8d70-00b3440d9127", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.508055] env[61868]: DEBUG nova.compute.manager [req-59b697cf-2489-440e-a9c0-caac34ddd5c5 req-6a71fcc5-feaa-4aa2-a21f-4201fbe6e3f2 service nova] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Received event network-vif-deleted-79fe6cfd-0c91-4ec5-95b2-22aed62a0e0f {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 914.572822] env[61868]: INFO nova.scheduler.client.report [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Deleted allocations for instance 083140bf-233d-49ac-8247-cd93206f9200 [ 914.573152] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3fc0b1d2-f149-4f6c-9662-ce7930d8b435 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Lock "083140bf-233d-49ac-8247-cd93206f9200" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 222.576s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.573939] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1f690bc9-2e92-4774-a822-25d8505bdac3 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Lock "083140bf-233d-49ac-8247-cd93206f9200" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 25.915s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.574315] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1f690bc9-2e92-4774-a822-25d8505bdac3 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Acquiring lock "083140bf-233d-49ac-8247-cd93206f9200-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.574610] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1f690bc9-2e92-4774-a822-25d8505bdac3 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Lock "083140bf-233d-49ac-8247-cd93206f9200-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.574805] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1f690bc9-2e92-4774-a822-25d8505bdac3 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Lock "083140bf-233d-49ac-8247-cd93206f9200-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.578588] env[61868]: INFO nova.compute.manager [None req-1f690bc9-2e92-4774-a822-25d8505bdac3 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Terminating instance [ 914.580354] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1f690bc9-2e92-4774-a822-25d8505bdac3 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Acquiring lock "refresh_cache-083140bf-233d-49ac-8247-cd93206f9200" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.580554] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1f690bc9-2e92-4774-a822-25d8505bdac3 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Acquired lock "refresh_cache-083140bf-233d-49ac-8247-cd93206f9200" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.580729] env[61868]: DEBUG nova.network.neutron [None req-1f690bc9-2e92-4774-a822-25d8505bdac3 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 914.681542] env[61868]: DEBUG nova.scheduler.client.report [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 914.773675] env[61868]: DEBUG nova.compute.manager [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 914.790850] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Releasing lock "refresh_cache-e59a182d-97b2-454a-bc40-8afb0839324f" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.791210] env[61868]: DEBUG nova.compute.manager [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Instance network_info: |[{"id": "cb19d64d-19c9-4d48-8d70-00b3440d9127", "address": "fa:16:3e:d4:cc:5c", "network": {"id": "3535e048-fe57-4cbc-9489-0b77484069bd", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-748518849-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c8440ce785c344d793dd513b5f008d65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb19d64d-19", "ovs_interfaceid": "cb19d64d-19c9-4d48-8d70-00b3440d9127", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 914.793161] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d4:cc:5c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'da0e5087-d65b-416f-90fe-beaa9c534ad3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cb19d64d-19c9-4d48-8d70-00b3440d9127', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 914.801284] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Creating folder: Project (c8440ce785c344d793dd513b5f008d65). Parent ref: group-v281478. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 914.807254] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-25cb3b73-e50e-4f57-9688-6ce2fe03a7c1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.816978] env[61868]: DEBUG nova.virt.hardware [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 914.817498] env[61868]: DEBUG nova.virt.hardware [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 914.817771] env[61868]: DEBUG nova.virt.hardware [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 914.818118] env[61868]: DEBUG nova.virt.hardware [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 914.818391] env[61868]: DEBUG nova.virt.hardware [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 914.818683] env[61868]: DEBUG nova.virt.hardware [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 914.819051] env[61868]: DEBUG nova.virt.hardware [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 914.821020] env[61868]: DEBUG nova.virt.hardware [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 914.821020] env[61868]: DEBUG nova.virt.hardware [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 914.821020] env[61868]: DEBUG nova.virt.hardware [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 914.821020] env[61868]: DEBUG nova.virt.hardware [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 914.822066] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82862e9a-d843-4776-9a65-f1eaa605e264 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.827543] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Created folder: Project (c8440ce785c344d793dd513b5f008d65) in parent group-v281478. [ 914.827859] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Creating folder: Instances. Parent ref: group-v281571. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 914.828612] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-09054109-23ac-405a-bb21-2197fe0eb257 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.835774] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67588a22-6254-4a3a-baef-a4ce8246e680 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.851432] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Instance VIF info [] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 914.857438] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Creating folder: Project (7acbf3c953424672b8f67cbc77b156ab). Parent ref: group-v281478. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 914.859099] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2632ba38-51bb-4fe4-9d37-8d149d844c7b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.860875] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Created folder: Instances in parent group-v281571. [ 914.861447] env[61868]: DEBUG oslo.service.loopingcall [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 914.861447] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 914.862048] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a0453d9f-b674-459f-9090-8d9c3f225c0e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.881107] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Created folder: Project (7acbf3c953424672b8f67cbc77b156ab) in parent group-v281478. [ 914.881107] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Creating folder: Instances. Parent ref: group-v281573. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 914.881107] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a26cdfc9-2e30-49f7-85a2-7b34c544de92 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.886074] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 914.886074] env[61868]: value = "task-1315649" [ 914.886074] env[61868]: _type = "Task" [ 914.886074] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.891580] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Created folder: Instances in parent group-v281573. [ 914.891824] env[61868]: DEBUG oslo.service.loopingcall [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 914.896279] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 914.896279] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315649, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.896279] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ebb64031-f1fc-400f-8356-cf02d8cde581 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.913750] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 914.913750] env[61868]: value = "task-1315651" [ 914.913750] env[61868]: _type = "Task" [ 914.913750] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.922223] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315651, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.003092] env[61868]: DEBUG nova.compute.manager [req-2f990086-6647-45ff-9fb1-cb493f6861a9 req-fb7b747d-83ff-4699-84e7-f79722226128 service nova] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Received event network-changed-cb19d64d-19c9-4d48-8d70-00b3440d9127 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 915.003595] env[61868]: DEBUG nova.compute.manager [req-2f990086-6647-45ff-9fb1-cb493f6861a9 req-fb7b747d-83ff-4699-84e7-f79722226128 service nova] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Refreshing instance network info cache due to event network-changed-cb19d64d-19c9-4d48-8d70-00b3440d9127. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 915.003595] env[61868]: DEBUG oslo_concurrency.lockutils [req-2f990086-6647-45ff-9fb1-cb493f6861a9 req-fb7b747d-83ff-4699-84e7-f79722226128 service nova] Acquiring lock "refresh_cache-e59a182d-97b2-454a-bc40-8afb0839324f" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.003595] env[61868]: DEBUG oslo_concurrency.lockutils [req-2f990086-6647-45ff-9fb1-cb493f6861a9 req-fb7b747d-83ff-4699-84e7-f79722226128 service nova] Acquired lock "refresh_cache-e59a182d-97b2-454a-bc40-8afb0839324f" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.003878] env[61868]: DEBUG nova.network.neutron [req-2f990086-6647-45ff-9fb1-cb493f6861a9 req-fb7b747d-83ff-4699-84e7-f79722226128 service nova] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Refreshing network info cache for port cb19d64d-19c9-4d48-8d70-00b3440d9127 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 915.105439] env[61868]: DEBUG nova.network.neutron [None req-1f690bc9-2e92-4774-a822-25d8505bdac3 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 915.187145] env[61868]: DEBUG oslo_concurrency.lockutils [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.439s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.187848] env[61868]: DEBUG nova.compute.manager [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 915.193904] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0f7f9081-032d-4547-939e-b812c9bdc0c7 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.497s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.198410] env[61868]: DEBUG nova.objects.instance [None req-0f7f9081-032d-4547-939e-b812c9bdc0c7 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Lazy-loading 'resources' on Instance uuid b9b5be37-6b30-4229-9c8d-3ee9d30db119 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 915.236760] env[61868]: DEBUG nova.network.neutron [None req-1f690bc9-2e92-4774-a822-25d8505bdac3 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.310396] env[61868]: INFO nova.compute.manager [None req-bc5ad1ea-ec84-4100-aa19-5bd2a6d097c9 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Unrescuing [ 915.313019] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bc5ad1ea-ec84-4100-aa19-5bd2a6d097c9 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Acquiring lock "refresh_cache-8ea8e28b-3582-45f2-b4b5-84f624415a58" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.313019] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bc5ad1ea-ec84-4100-aa19-5bd2a6d097c9 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Acquired lock "refresh_cache-8ea8e28b-3582-45f2-b4b5-84f624415a58" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.313019] env[61868]: DEBUG nova.network.neutron [None req-bc5ad1ea-ec84-4100-aa19-5bd2a6d097c9 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 915.313857] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-001b27d5-1800-43bb-b6d4-35be0ea48f57 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Volume attach. Driver type: vmdk {{(pid=61868) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 915.314270] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-001b27d5-1800-43bb-b6d4-35be0ea48f57 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281570', 'volume_id': 'fdcd4a1e-341a-4607-b405-3e3c2a834712', 'name': 'volume-fdcd4a1e-341a-4607-b405-3e3c2a834712', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '409cad1a-946d-4c58-aa57-1c0bf97fe63f', 'attached_at': '', 'detached_at': '', 'volume_id': 'fdcd4a1e-341a-4607-b405-3e3c2a834712', 'serial': 'fdcd4a1e-341a-4607-b405-3e3c2a834712'} {{(pid=61868) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 915.315353] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b99d429c-e66d-463c-a300-2c85554c8d74 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.344433] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20e12b81-8ced-49a9-8f70-64b0eacfe397 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.352994] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Acquiring lock "87b25494-1853-4c7b-ae8b-6283992e895c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.353627] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Lock "87b25494-1853-4c7b-ae8b-6283992e895c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.383435] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-001b27d5-1800-43bb-b6d4-35be0ea48f57 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] volume-fdcd4a1e-341a-4607-b405-3e3c2a834712/volume-fdcd4a1e-341a-4607-b405-3e3c2a834712.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 915.384411] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-50fb149e-89a2-4faf-9ffb-97be2e9141c7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.408811] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315649, 'name': CreateVM_Task, 'duration_secs': 0.442095} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.410015] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 915.410316] env[61868]: DEBUG oslo_vmware.api [None req-001b27d5-1800-43bb-b6d4-35be0ea48f57 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Waiting for the task: (returnval){ [ 915.410316] env[61868]: value = "task-1315652" [ 915.410316] env[61868]: _type = "Task" [ 915.410316] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.410938] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.411120] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.411436] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 915.411734] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e5bc249-e5c8-4aab-bc9d-734b2575d1bd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.420421] env[61868]: DEBUG oslo_vmware.api [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 915.420421] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5231c7b0-f6d0-e7cf-0cb4-ec881622e0ec" [ 915.420421] env[61868]: _type = "Task" [ 915.420421] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.431407] env[61868]: DEBUG oslo_vmware.api [None req-001b27d5-1800-43bb-b6d4-35be0ea48f57 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315652, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.431622] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315651, 'name': CreateVM_Task, 'duration_secs': 0.383537} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.432194] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 915.432917] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.437012] env[61868]: DEBUG oslo_vmware.api [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5231c7b0-f6d0-e7cf-0cb4-ec881622e0ec, 'name': SearchDatastore_Task, 'duration_secs': 0.012138} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.437333] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.437543] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 915.437870] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.437922] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.438084] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 915.438393] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.438702] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 915.438922] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9c016fdc-e2f5-41e2-b54c-d872a1a0c35d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.440852] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c673c5d2-04a2-4a12-99c4-e5085ebb4272 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.445911] env[61868]: DEBUG oslo_vmware.api [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Waiting for the task: (returnval){ [ 915.445911] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5258b5be-3be3-40d2-4c0c-5addc81d45b0" [ 915.445911] env[61868]: _type = "Task" [ 915.445911] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.449997] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 915.450259] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 915.451237] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e30d977-5d28-4904-b664-3d2abebe5e78 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.456955] env[61868]: DEBUG oslo_vmware.api [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5258b5be-3be3-40d2-4c0c-5addc81d45b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.459815] env[61868]: DEBUG oslo_vmware.api [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 915.459815] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52cfb86a-2942-a334-3740-6df41914ec24" [ 915.459815] env[61868]: _type = "Task" [ 915.459815] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.468137] env[61868]: DEBUG oslo_vmware.api [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52cfb86a-2942-a334-3740-6df41914ec24, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.679612] env[61868]: DEBUG oslo_vmware.rw_handles [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ddacd5-ae46-34f7-3c76-1835835a4316/disk-0.vmdk. {{(pid=61868) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 915.680544] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f960c3d-fb31-41df-b2a4-f49384efab61 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.687098] env[61868]: DEBUG oslo_vmware.rw_handles [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ddacd5-ae46-34f7-3c76-1835835a4316/disk-0.vmdk is in state: ready. {{(pid=61868) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 915.687299] env[61868]: ERROR oslo_vmware.rw_handles [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ddacd5-ae46-34f7-3c76-1835835a4316/disk-0.vmdk due to incomplete transfer. [ 915.687555] env[61868]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-7a0b424c-4ca2-4ada-8542-3a86984f7fd8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.695835] env[61868]: DEBUG oslo_vmware.rw_handles [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ddacd5-ae46-34f7-3c76-1835835a4316/disk-0.vmdk. {{(pid=61868) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 915.696059] env[61868]: DEBUG nova.virt.vmwareapi.images [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Uploaded image ba384da7-fddc-4e13-bac6-b4e3d1857675 to the Glance image server {{(pid=61868) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 915.698792] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Destroying the VM {{(pid=61868) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 915.698957] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-5105a927-cf17-4e49-a265-422ea05aa523 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.702394] env[61868]: DEBUG nova.compute.utils [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 915.706182] env[61868]: DEBUG nova.compute.manager [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 915.706357] env[61868]: DEBUG nova.network.neutron [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 915.715162] env[61868]: DEBUG oslo_vmware.api [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 915.715162] env[61868]: value = "task-1315653" [ 915.715162] env[61868]: _type = "Task" [ 915.715162] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.723211] env[61868]: DEBUG oslo_vmware.api [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315653, 'name': Destroy_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.734966] env[61868]: DEBUG nova.network.neutron [req-2f990086-6647-45ff-9fb1-cb493f6861a9 req-fb7b747d-83ff-4699-84e7-f79722226128 service nova] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Updated VIF entry in instance network info cache for port cb19d64d-19c9-4d48-8d70-00b3440d9127. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 915.735435] env[61868]: DEBUG nova.network.neutron [req-2f990086-6647-45ff-9fb1-cb493f6861a9 req-fb7b747d-83ff-4699-84e7-f79722226128 service nova] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Updating instance_info_cache with network_info: [{"id": "cb19d64d-19c9-4d48-8d70-00b3440d9127", "address": "fa:16:3e:d4:cc:5c", "network": {"id": "3535e048-fe57-4cbc-9489-0b77484069bd", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-748518849-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c8440ce785c344d793dd513b5f008d65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb19d64d-19", "ovs_interfaceid": "cb19d64d-19c9-4d48-8d70-00b3440d9127", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.741478] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1f690bc9-2e92-4774-a822-25d8505bdac3 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Releasing lock "refresh_cache-083140bf-233d-49ac-8247-cd93206f9200" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.741968] env[61868]: DEBUG nova.compute.manager [None req-1f690bc9-2e92-4774-a822-25d8505bdac3 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 915.742216] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1f690bc9-2e92-4774-a822-25d8505bdac3 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 915.743142] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2dc78da-b0df-44a9-9ddf-a8b866be819b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.754067] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f690bc9-2e92-4774-a822-25d8505bdac3 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 915.754357] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c83b9b74-f429-43a5-a6b4-695e06cddc1c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.761768] env[61868]: DEBUG nova.policy [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fd6ed09d33f6491b83dc72cef31bfd29', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '812e115cc1114d0e8536924c4f5556f8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 915.764295] env[61868]: DEBUG oslo_vmware.api [None req-1f690bc9-2e92-4774-a822-25d8505bdac3 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Waiting for the task: (returnval){ [ 915.764295] env[61868]: value = "task-1315654" [ 915.764295] env[61868]: _type = "Task" [ 915.764295] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.774635] env[61868]: DEBUG oslo_vmware.api [None req-1f690bc9-2e92-4774-a822-25d8505bdac3 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Task: {'id': task-1315654, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.856453] env[61868]: DEBUG nova.compute.manager [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 915.924596] env[61868]: DEBUG oslo_vmware.api [None req-001b27d5-1800-43bb-b6d4-35be0ea48f57 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315652, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.958687] env[61868]: DEBUG oslo_vmware.api [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5258b5be-3be3-40d2-4c0c-5addc81d45b0, 'name': SearchDatastore_Task, 'duration_secs': 0.011097} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.961936] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.962240] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 915.962497] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.974106] env[61868]: DEBUG oslo_vmware.api [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52cfb86a-2942-a334-3740-6df41914ec24, 'name': SearchDatastore_Task, 'duration_secs': 0.010365} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.975055] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4997aeb-ee5e-4080-837a-c016cc2b18fc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.988026] env[61868]: DEBUG oslo_vmware.api [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 915.988026] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]522ff813-297d-535a-e3f9-74ae4ba738bd" [ 915.988026] env[61868]: _type = "Task" [ 915.988026] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.999144] env[61868]: DEBUG oslo_vmware.api [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]522ff813-297d-535a-e3f9-74ae4ba738bd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.020628] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cc65862-91ef-4e85-aab7-e30be02a3eaa {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.031477] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a14fe265-56c0-4499-8d0c-8d8299345623 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.836310] env[61868]: DEBUG nova.compute.manager [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 916.836913] env[61868]: DEBUG oslo_concurrency.lockutils [req-2f990086-6647-45ff-9fb1-cb493f6861a9 req-fb7b747d-83ff-4699-84e7-f79722226128 service nova] Releasing lock "refresh_cache-e59a182d-97b2-454a-bc40-8afb0839324f" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.841702] env[61868]: DEBUG nova.network.neutron [None req-bc5ad1ea-ec84-4100-aa19-5bd2a6d097c9 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Updating instance_info_cache with network_info: [{"id": "c27d3e06-cf3c-4ef8-ba0b-e6f2ca824646", "address": "fa:16:3e:3e:8f:76", "network": {"id": "094eaf15-c932-4d4f-8736-fe5239160507", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1905020504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "7541757d43c74b93acf728aa2fb0f425", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc27d3e06-cf", "ovs_interfaceid": "c27d3e06-cf3c-4ef8-ba0b-e6f2ca824646", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.849242] env[61868]: DEBUG nova.network.neutron [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Successfully created port: ec859835-c58b-4c26-ac77-7628f73990b0 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 916.860774] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3248a9b2-d268-4c54-8529-5eb5f5d0dc54 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.872190] env[61868]: DEBUG oslo_vmware.api [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315653, 'name': Destroy_Task, 'duration_secs': 0.751998} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.884065] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Destroyed the VM [ 916.884365] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Deleting Snapshot of the VM instance {{(pid=61868) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 916.884702] env[61868]: DEBUG oslo_vmware.api [None req-001b27d5-1800-43bb-b6d4-35be0ea48f57 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315652, 'name': ReconfigVM_Task, 'duration_secs': 0.907746} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.885702] env[61868]: DEBUG oslo_vmware.api [None req-1f690bc9-2e92-4774-a822-25d8505bdac3 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Task: {'id': task-1315654, 'name': PowerOffVM_Task, 'duration_secs': 0.54589} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.885913] env[61868]: DEBUG oslo_vmware.api [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]522ff813-297d-535a-e3f9-74ae4ba738bd, 'name': SearchDatastore_Task, 'duration_secs': 0.012193} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.887370] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.887617] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-a560c6cb-44bf-445c-8e73-b2775d944e57 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.889454] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-001b27d5-1800-43bb-b6d4-35be0ea48f57 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Reconfigured VM instance instance-0000003f to attach disk [datastore2] volume-fdcd4a1e-341a-4607-b405-3e3c2a834712/volume-fdcd4a1e-341a-4607-b405-3e3c2a834712.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 916.894115] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f690bc9-2e92-4774-a822-25d8505bdac3 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 916.894301] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1f690bc9-2e92-4774-a822-25d8505bdac3 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 916.894596] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.894842] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] e59a182d-97b2-454a-bc40-8afb0839324f/e59a182d-97b2-454a-bc40-8afb0839324f.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 916.897981] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ad76c315-c32c-4fc3-b6d0-1aab1e4cf632 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.907662] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a49a5607-ca0a-4571-b920-6c1ceb030470 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.909192] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.909414] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 916.909646] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dda82b99-116a-48ec-9e7a-3f1a0fc70d37 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.912460] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a739bdf0-77d5-49db-a3df-e63fa8db1a5d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.915117] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c478e33d-cf64-4350-b7fa-702ef74ee9c2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.921728] env[61868]: DEBUG oslo_vmware.api [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 916.921728] env[61868]: value = "task-1315655" [ 916.921728] env[61868]: _type = "Task" [ 916.921728] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.932060] env[61868]: DEBUG oslo_vmware.api [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 916.932060] env[61868]: value = "task-1315658" [ 916.932060] env[61868]: _type = "Task" [ 916.932060] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.932060] env[61868]: DEBUG oslo_vmware.api [None req-001b27d5-1800-43bb-b6d4-35be0ea48f57 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Waiting for the task: (returnval){ [ 916.932060] env[61868]: value = "task-1315657" [ 916.932060] env[61868]: _type = "Task" [ 916.932060] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.950204] env[61868]: DEBUG nova.compute.provider_tree [None req-0f7f9081-032d-4547-939e-b812c9bdc0c7 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 916.951626] env[61868]: DEBUG oslo_vmware.api [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315655, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.951873] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 916.952081] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 916.953494] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c16da8e-c422-4677-a6fd-cf6f1db25c16 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.962490] env[61868]: DEBUG oslo_vmware.api [None req-001b27d5-1800-43bb-b6d4-35be0ea48f57 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315657, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.962641] env[61868]: DEBUG oslo_vmware.api [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315658, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.968693] env[61868]: DEBUG oslo_vmware.api [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Waiting for the task: (returnval){ [ 916.968693] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]529701a8-2e3c-bdc1-1ca5-33fd023c4d10" [ 916.968693] env[61868]: _type = "Task" [ 916.968693] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.975991] env[61868]: DEBUG oslo_vmware.api [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]529701a8-2e3c-bdc1-1ca5-33fd023c4d10, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.994649] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1f690bc9-2e92-4774-a822-25d8505bdac3 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 916.994922] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1f690bc9-2e92-4774-a822-25d8505bdac3 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 916.995132] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f690bc9-2e92-4774-a822-25d8505bdac3 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Deleting the datastore file [datastore1] 083140bf-233d-49ac-8247-cd93206f9200 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 916.995405] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3b4338fd-e063-4a46-b357-f4901acd95df {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.004245] env[61868]: DEBUG oslo_vmware.api [None req-1f690bc9-2e92-4774-a822-25d8505bdac3 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Waiting for the task: (returnval){ [ 917.004245] env[61868]: value = "task-1315659" [ 917.004245] env[61868]: _type = "Task" [ 917.004245] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.012777] env[61868]: DEBUG oslo_vmware.api [None req-1f690bc9-2e92-4774-a822-25d8505bdac3 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Task: {'id': task-1315659, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.365807] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e5888991-048d-4aa8-b4bd-2d8f6de4700f tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "306f2dc4-933c-4948-8330-eda93931c4e0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.366082] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e5888991-048d-4aa8-b4bd-2d8f6de4700f tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "306f2dc4-933c-4948-8330-eda93931c4e0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.366304] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e5888991-048d-4aa8-b4bd-2d8f6de4700f tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "306f2dc4-933c-4948-8330-eda93931c4e0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.366511] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e5888991-048d-4aa8-b4bd-2d8f6de4700f tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "306f2dc4-933c-4948-8330-eda93931c4e0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.366684] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e5888991-048d-4aa8-b4bd-2d8f6de4700f tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "306f2dc4-933c-4948-8330-eda93931c4e0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.368819] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bc5ad1ea-ec84-4100-aa19-5bd2a6d097c9 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Releasing lock "refresh_cache-8ea8e28b-3582-45f2-b4b5-84f624415a58" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.369394] env[61868]: DEBUG nova.objects.instance [None req-bc5ad1ea-ec84-4100-aa19-5bd2a6d097c9 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Lazy-loading 'flavor' on Instance uuid 8ea8e28b-3582-45f2-b4b5-84f624415a58 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 917.370915] env[61868]: INFO nova.compute.manager [None req-e5888991-048d-4aa8-b4bd-2d8f6de4700f tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Terminating instance [ 917.373830] env[61868]: DEBUG nova.compute.manager [None req-e5888991-048d-4aa8-b4bd-2d8f6de4700f tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 917.374079] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-e5888991-048d-4aa8-b4bd-2d8f6de4700f tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 917.375173] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b754bbd-444b-41c3-839a-0aaece5e7cce {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.386605] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-e5888991-048d-4aa8-b4bd-2d8f6de4700f tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 917.386883] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3cde1911-ad79-4039-b729-1459234099a5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.436190] env[61868]: DEBUG oslo_vmware.api [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315655, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.449405] env[61868]: DEBUG oslo_vmware.api [None req-001b27d5-1800-43bb-b6d4-35be0ea48f57 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315657, 'name': ReconfigVM_Task, 'duration_secs': 0.20321} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.453116] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-001b27d5-1800-43bb-b6d4-35be0ea48f57 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281570', 'volume_id': 'fdcd4a1e-341a-4607-b405-3e3c2a834712', 'name': 'volume-fdcd4a1e-341a-4607-b405-3e3c2a834712', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '409cad1a-946d-4c58-aa57-1c0bf97fe63f', 'attached_at': '', 'detached_at': '', 'volume_id': 'fdcd4a1e-341a-4607-b405-3e3c2a834712', 'serial': 'fdcd4a1e-341a-4607-b405-3e3c2a834712'} {{(pid=61868) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 917.454883] env[61868]: DEBUG oslo_vmware.api [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315658, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.460209] env[61868]: DEBUG nova.scheduler.client.report [None req-0f7f9081-032d-4547-939e-b812c9bdc0c7 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 917.479604] env[61868]: DEBUG oslo_vmware.api [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]529701a8-2e3c-bdc1-1ca5-33fd023c4d10, 'name': SearchDatastore_Task, 'duration_secs': 0.01163} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.483474] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10363a16-b0ce-44c0-8f81-526063755103 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.488326] env[61868]: DEBUG oslo_vmware.api [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Waiting for the task: (returnval){ [ 917.488326] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e7e1d1-2c65-71e0-e503-2179e32c8675" [ 917.488326] env[61868]: _type = "Task" [ 917.488326] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.500079] env[61868]: DEBUG oslo_vmware.api [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e7e1d1-2c65-71e0-e503-2179e32c8675, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.514218] env[61868]: DEBUG oslo_vmware.api [None req-1f690bc9-2e92-4774-a822-25d8505bdac3 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Task: {'id': task-1315659, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.320127} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.514543] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f690bc9-2e92-4774-a822-25d8505bdac3 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 917.514758] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1f690bc9-2e92-4774-a822-25d8505bdac3 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 917.514963] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1f690bc9-2e92-4774-a822-25d8505bdac3 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 917.515161] env[61868]: INFO nova.compute.manager [None req-1f690bc9-2e92-4774-a822-25d8505bdac3 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Took 1.77 seconds to destroy the instance on the hypervisor. [ 917.515420] env[61868]: DEBUG oslo.service.loopingcall [None req-1f690bc9-2e92-4774-a822-25d8505bdac3 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 917.515621] env[61868]: DEBUG nova.compute.manager [-] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 917.515714] env[61868]: DEBUG nova.network.neutron [-] [instance: 083140bf-233d-49ac-8247-cd93206f9200] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 917.534699] env[61868]: DEBUG nova.network.neutron [-] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 917.558925] env[61868]: DEBUG oslo_concurrency.lockutils [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "867384c6-175b-4848-b9ac-917470742ba2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.559196] env[61868]: DEBUG oslo_concurrency.lockutils [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "867384c6-175b-4848-b9ac-917470742ba2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.848570] env[61868]: DEBUG nova.compute.manager [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 917.873482] env[61868]: DEBUG nova.virt.hardware [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 917.873747] env[61868]: DEBUG nova.virt.hardware [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 917.873907] env[61868]: DEBUG nova.virt.hardware [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 917.874115] env[61868]: DEBUG nova.virt.hardware [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 917.874260] env[61868]: DEBUG nova.virt.hardware [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 917.874408] env[61868]: DEBUG nova.virt.hardware [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 917.874618] env[61868]: DEBUG nova.virt.hardware [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 917.874777] env[61868]: DEBUG nova.virt.hardware [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 917.874998] env[61868]: DEBUG nova.virt.hardware [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 917.875141] env[61868]: DEBUG nova.virt.hardware [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 917.875313] env[61868]: DEBUG nova.virt.hardware [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 917.878058] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40bc67e7-5ef9-45c2-b7c6-ff0d92fc8fef {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.881502] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4905cbb-ae9f-4296-becb-2cfa56419570 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.905145] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58df331e-2439-48ae-be73-98ca7ed9be2c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.909224] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc5ad1ea-ec84-4100-aa19-5bd2a6d097c9 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 917.909563] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d8d2f707-383f-4db2-960d-ac80af912006 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.922463] env[61868]: DEBUG oslo_vmware.api [None req-bc5ad1ea-ec84-4100-aa19-5bd2a6d097c9 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 917.922463] env[61868]: value = "task-1315661" [ 917.922463] env[61868]: _type = "Task" [ 917.922463] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.933518] env[61868]: DEBUG oslo_vmware.api [None req-bc5ad1ea-ec84-4100-aa19-5bd2a6d097c9 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315661, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.938202] env[61868]: DEBUG oslo_vmware.api [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315655, 'name': RemoveSnapshot_Task, 'duration_secs': 0.792232} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.944471] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Deleted Snapshot of the VM instance {{(pid=61868) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 917.944737] env[61868]: INFO nova.compute.manager [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Took 17.37 seconds to snapshot the instance on the hypervisor. [ 917.947538] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-e5888991-048d-4aa8-b4bd-2d8f6de4700f tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 917.947772] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-e5888991-048d-4aa8-b4bd-2d8f6de4700f tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 917.948022] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5888991-048d-4aa8-b4bd-2d8f6de4700f tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Deleting the datastore file [datastore1] 306f2dc4-933c-4948-8330-eda93931c4e0 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 917.948513] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-815c9090-196a-4b69-bf0d-6f533ed74f10 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.958842] env[61868]: DEBUG oslo_vmware.api [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315658, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.763085} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.961703] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] e59a182d-97b2-454a-bc40-8afb0839324f/e59a182d-97b2-454a-bc40-8afb0839324f.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 917.961996] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 917.962924] env[61868]: DEBUG oslo_vmware.api [None req-e5888991-048d-4aa8-b4bd-2d8f6de4700f tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 917.962924] env[61868]: value = "task-1315662" [ 917.962924] env[61868]: _type = "Task" [ 917.962924] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.962924] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9ac55429-3da2-4bac-92c3-8fa1cabb6a19 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.970769] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0f7f9081-032d-4547-939e-b812c9bdc0c7 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.777s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.974452] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4dcc6a68-7a7e-4a85-8566-794266e676cd tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.537s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.974721] env[61868]: DEBUG nova.objects.instance [None req-4dcc6a68-7a7e-4a85-8566-794266e676cd tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Lazy-loading 'resources' on Instance uuid c6108cf2-b597-4ca7-8f57-12ea82cc6c2c {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 917.976622] env[61868]: DEBUG oslo_vmware.api [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 917.976622] env[61868]: value = "task-1315663" [ 917.976622] env[61868]: _type = "Task" [ 917.976622] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.984050] env[61868]: DEBUG oslo_vmware.api [None req-e5888991-048d-4aa8-b4bd-2d8f6de4700f tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315662, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.990021] env[61868]: DEBUG oslo_vmware.api [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315663, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.999968] env[61868]: DEBUG oslo_vmware.api [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e7e1d1-2c65-71e0-e503-2179e32c8675, 'name': SearchDatastore_Task, 'duration_secs': 0.058774} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.000261] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.000519] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] b51e968e-6911-47f3-b5a9-e8a5865e42a0/b51e968e-6911-47f3-b5a9-e8a5865e42a0.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 918.000823] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-18c0f907-b062-44a5-90c3-1ba02cb637af {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.006308] env[61868]: INFO nova.scheduler.client.report [None req-0f7f9081-032d-4547-939e-b812c9bdc0c7 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Deleted allocations for instance b9b5be37-6b30-4229-9c8d-3ee9d30db119 [ 918.008542] env[61868]: DEBUG oslo_vmware.api [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Waiting for the task: (returnval){ [ 918.008542] env[61868]: value = "task-1315664" [ 918.008542] env[61868]: _type = "Task" [ 918.008542] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.025481] env[61868]: DEBUG oslo_vmware.api [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': task-1315664, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.037573] env[61868]: DEBUG nova.network.neutron [-] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.061637] env[61868]: DEBUG nova.compute.manager [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 918.439537] env[61868]: DEBUG oslo_vmware.api [None req-bc5ad1ea-ec84-4100-aa19-5bd2a6d097c9 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315661, 'name': PowerOffVM_Task, 'duration_secs': 0.229214} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.439896] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc5ad1ea-ec84-4100-aa19-5bd2a6d097c9 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 918.445792] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc5ad1ea-ec84-4100-aa19-5bd2a6d097c9 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Reconfiguring VM instance instance-0000004d to detach disk 2001 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 918.446164] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a170bec6-7639-4328-944a-78d6ce972dd8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.460049] env[61868]: DEBUG oslo_concurrency.lockutils [None req-fc9e1fb6-3a19-4eb7-b4e6-935739b166e1 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "interface-5aec2609-79d3-4725-a182-70b575adbe17-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.460334] env[61868]: DEBUG oslo_concurrency.lockutils [None req-fc9e1fb6-3a19-4eb7-b4e6-935739b166e1 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "interface-5aec2609-79d3-4725-a182-70b575adbe17-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.460708] env[61868]: DEBUG nova.objects.instance [None req-fc9e1fb6-3a19-4eb7-b4e6-935739b166e1 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lazy-loading 'flavor' on Instance uuid 5aec2609-79d3-4725-a182-70b575adbe17 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 918.462301] env[61868]: DEBUG nova.compute.manager [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Instance disappeared during snapshot {{(pid=61868) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 918.472196] env[61868]: DEBUG oslo_vmware.api [None req-bc5ad1ea-ec84-4100-aa19-5bd2a6d097c9 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 918.472196] env[61868]: value = "task-1315665" [ 918.472196] env[61868]: _type = "Task" [ 918.472196] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.482813] env[61868]: DEBUG oslo_vmware.api [None req-e5888991-048d-4aa8-b4bd-2d8f6de4700f tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315662, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.207456} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.487540] env[61868]: DEBUG nova.compute.manager [None req-ba2b1f0b-5ec0-44d1-9946-7014aeda035c tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Image not found during clean up ba384da7-fddc-4e13-bac6-b4e3d1857675 {{(pid=61868) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4500}} [ 918.492369] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5888991-048d-4aa8-b4bd-2d8f6de4700f tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 918.492681] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-e5888991-048d-4aa8-b4bd-2d8f6de4700f tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 918.492828] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-e5888991-048d-4aa8-b4bd-2d8f6de4700f tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 918.493023] env[61868]: INFO nova.compute.manager [None req-e5888991-048d-4aa8-b4bd-2d8f6de4700f tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Took 1.12 seconds to destroy the instance on the hypervisor. [ 918.493297] env[61868]: DEBUG oslo.service.loopingcall [None req-e5888991-048d-4aa8-b4bd-2d8f6de4700f tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 918.493860] env[61868]: DEBUG oslo_vmware.api [None req-bc5ad1ea-ec84-4100-aa19-5bd2a6d097c9 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315665, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.494436] env[61868]: DEBUG nova.compute.manager [-] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 918.494551] env[61868]: DEBUG nova.network.neutron [-] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 918.503480] env[61868]: DEBUG oslo_vmware.api [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315663, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067854} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.503820] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 918.504715] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d9b990e-adf3-42e1-aae6-d1b85b8cdf08 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.532785] env[61868]: DEBUG nova.objects.instance [None req-001b27d5-1800-43bb-b6d4-35be0ea48f57 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Lazy-loading 'flavor' on Instance uuid 409cad1a-946d-4c58-aa57-1c0bf97fe63f {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 918.544476] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] e59a182d-97b2-454a-bc40-8afb0839324f/e59a182d-97b2-454a-bc40-8afb0839324f.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 918.548024] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0f7f9081-032d-4547-939e-b812c9bdc0c7 tempest-VolumesAdminNegativeTest-1488838211 tempest-VolumesAdminNegativeTest-1488838211-project-member] Lock "b9b5be37-6b30-4229-9c8d-3ee9d30db119" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.094s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.549459] env[61868]: INFO nova.compute.manager [-] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Took 1.03 seconds to deallocate network for instance. [ 918.549713] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d7094069-f866-47dc-ab9e-d6465e1e8257 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.576647] env[61868]: DEBUG oslo_vmware.api [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': task-1315664, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.541711} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.579147] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] b51e968e-6911-47f3-b5a9-e8a5865e42a0/b51e968e-6911-47f3-b5a9-e8a5865e42a0.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 918.579411] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 918.583962] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1d47153a-7640-425c-a5e1-40ffaa20633e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.585515] env[61868]: DEBUG oslo_vmware.api [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 918.585515] env[61868]: value = "task-1315666" [ 918.585515] env[61868]: _type = "Task" [ 918.585515] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.590934] env[61868]: DEBUG oslo_vmware.api [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Waiting for the task: (returnval){ [ 918.590934] env[61868]: value = "task-1315667" [ 918.590934] env[61868]: _type = "Task" [ 918.590934] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.602540] env[61868]: DEBUG oslo_vmware.api [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315666, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.605796] env[61868]: DEBUG oslo_concurrency.lockutils [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.610721] env[61868]: DEBUG oslo_vmware.api [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': task-1315667, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.800958] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a960a49-3703-4e1e-a6c4-a711d8c514ba {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.810482] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf1c123e-0993-45e7-bef0-12674918ccdd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.848511] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d38da1e5-3191-4288-a852-c4569cfae20c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.857015] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a67fe12-dcc5-4bc6-90a5-213f72812036 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.871758] env[61868]: DEBUG nova.compute.provider_tree [None req-4dcc6a68-7a7e-4a85-8566-794266e676cd tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 918.966750] env[61868]: DEBUG nova.objects.instance [None req-fc9e1fb6-3a19-4eb7-b4e6-935739b166e1 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lazy-loading 'pci_requests' on Instance uuid 5aec2609-79d3-4725-a182-70b575adbe17 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 918.983053] env[61868]: DEBUG oslo_vmware.api [None req-bc5ad1ea-ec84-4100-aa19-5bd2a6d097c9 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315665, 'name': ReconfigVM_Task, 'duration_secs': 0.263104} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.983328] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc5ad1ea-ec84-4100-aa19-5bd2a6d097c9 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Reconfigured VM instance instance-0000004d to detach disk 2001 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 918.983519] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc5ad1ea-ec84-4100-aa19-5bd2a6d097c9 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 918.983759] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c8007cce-fefd-4ff4-8cc3-52053694ae2d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.989693] env[61868]: DEBUG oslo_vmware.api [None req-bc5ad1ea-ec84-4100-aa19-5bd2a6d097c9 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 918.989693] env[61868]: value = "task-1315668" [ 918.989693] env[61868]: _type = "Task" [ 918.989693] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.997690] env[61868]: DEBUG oslo_vmware.api [None req-bc5ad1ea-ec84-4100-aa19-5bd2a6d097c9 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315668, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.033614] env[61868]: DEBUG nova.network.neutron [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Successfully updated port: ec859835-c58b-4c26-ac77-7628f73990b0 {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 919.052883] env[61868]: DEBUG oslo_concurrency.lockutils [None req-001b27d5-1800-43bb-b6d4-35be0ea48f57 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Lock "409cad1a-946d-4c58-aa57-1c0bf97fe63f" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.388s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.079194] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1f690bc9-2e92-4774-a822-25d8505bdac3 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.094942] env[61868]: DEBUG oslo_vmware.api [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315666, 'name': ReconfigVM_Task, 'duration_secs': 0.291196} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.099707] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Reconfigured VM instance instance-0000004f to attach disk [datastore2] e59a182d-97b2-454a-bc40-8afb0839324f/e59a182d-97b2-454a-bc40-8afb0839324f.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 919.100577] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e8b4489a-4bfa-4993-a422-e3c9ac18a167 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.106758] env[61868]: DEBUG oslo_vmware.api [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': task-1315667, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075608} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.107907] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 919.108266] env[61868]: DEBUG oslo_vmware.api [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 919.108266] env[61868]: value = "task-1315669" [ 919.108266] env[61868]: _type = "Task" [ 919.108266] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.108956] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b39558c-8785-418d-b930-ac59f0f241e7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.115876] env[61868]: DEBUG nova.compute.manager [req-5f444c96-a62d-4762-b643-e207e84fd349 req-6eee7328-914b-436a-b6ea-0ddbb25eee79 service nova] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Received event network-vif-deleted-4186aca7-63b5-4b24-ab7c-de8f87064242 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 919.116101] env[61868]: INFO nova.compute.manager [req-5f444c96-a62d-4762-b643-e207e84fd349 req-6eee7328-914b-436a-b6ea-0ddbb25eee79 service nova] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Neutron deleted interface 4186aca7-63b5-4b24-ab7c-de8f87064242; detaching it from the instance and deleting it from the info cache [ 919.116299] env[61868]: DEBUG nova.network.neutron [req-5f444c96-a62d-4762-b643-e207e84fd349 req-6eee7328-914b-436a-b6ea-0ddbb25eee79 service nova] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.134417] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] b51e968e-6911-47f3-b5a9-e8a5865e42a0/b51e968e-6911-47f3-b5a9-e8a5865e42a0.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 919.139858] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-261a5022-63e7-457d-9972-8ffd886fd3af {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.153995] env[61868]: DEBUG nova.compute.manager [req-5d3059aa-e7d3-4b8e-a39d-0907db1e6192 req-19d49be0-cd59-4ad3-a5b6-2d65e816fded service nova] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Received event network-vif-plugged-ec859835-c58b-4c26-ac77-7628f73990b0 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 919.154303] env[61868]: DEBUG oslo_concurrency.lockutils [req-5d3059aa-e7d3-4b8e-a39d-0907db1e6192 req-19d49be0-cd59-4ad3-a5b6-2d65e816fded service nova] Acquiring lock "4e9d75e8-ca33-4e33-baa3-cb7d596f584e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.154541] env[61868]: DEBUG oslo_concurrency.lockutils [req-5d3059aa-e7d3-4b8e-a39d-0907db1e6192 req-19d49be0-cd59-4ad3-a5b6-2d65e816fded service nova] Lock "4e9d75e8-ca33-4e33-baa3-cb7d596f584e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.154750] env[61868]: DEBUG oslo_concurrency.lockutils [req-5d3059aa-e7d3-4b8e-a39d-0907db1e6192 req-19d49be0-cd59-4ad3-a5b6-2d65e816fded service nova] Lock "4e9d75e8-ca33-4e33-baa3-cb7d596f584e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.154935] env[61868]: DEBUG nova.compute.manager [req-5d3059aa-e7d3-4b8e-a39d-0907db1e6192 req-19d49be0-cd59-4ad3-a5b6-2d65e816fded service nova] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] No waiting events found dispatching network-vif-plugged-ec859835-c58b-4c26-ac77-7628f73990b0 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 919.155140] env[61868]: WARNING nova.compute.manager [req-5d3059aa-e7d3-4b8e-a39d-0907db1e6192 req-19d49be0-cd59-4ad3-a5b6-2d65e816fded service nova] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Received unexpected event network-vif-plugged-ec859835-c58b-4c26-ac77-7628f73990b0 for instance with vm_state building and task_state spawning. [ 919.155895] env[61868]: DEBUG oslo_vmware.api [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315669, 'name': Rename_Task} progress is 10%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.161071] env[61868]: DEBUG oslo_vmware.api [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Waiting for the task: (returnval){ [ 919.161071] env[61868]: value = "task-1315670" [ 919.161071] env[61868]: _type = "Task" [ 919.161071] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.168942] env[61868]: DEBUG oslo_vmware.api [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': task-1315670, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.317275] env[61868]: DEBUG nova.network.neutron [-] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.375092] env[61868]: DEBUG nova.scheduler.client.report [None req-4dcc6a68-7a7e-4a85-8566-794266e676cd tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 919.469438] env[61868]: DEBUG nova.objects.base [None req-fc9e1fb6-3a19-4eb7-b4e6-935739b166e1 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Object Instance<5aec2609-79d3-4725-a182-70b575adbe17> lazy-loaded attributes: flavor,pci_requests {{(pid=61868) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 919.469682] env[61868]: DEBUG nova.network.neutron [None req-fc9e1fb6-3a19-4eb7-b4e6-935739b166e1 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 919.501785] env[61868]: DEBUG oslo_vmware.api [None req-bc5ad1ea-ec84-4100-aa19-5bd2a6d097c9 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315668, 'name': PowerOnVM_Task, 'duration_secs': 0.387175} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.502139] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc5ad1ea-ec84-4100-aa19-5bd2a6d097c9 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 919.502388] env[61868]: DEBUG nova.compute.manager [None req-bc5ad1ea-ec84-4100-aa19-5bd2a6d097c9 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 919.503246] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad1f594c-2e9a-4ef3-84d7-195a57bf5612 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.536917] env[61868]: DEBUG oslo_concurrency.lockutils [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "refresh_cache-4e9d75e8-ca33-4e33-baa3-cb7d596f584e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.537086] env[61868]: DEBUG oslo_concurrency.lockutils [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquired lock "refresh_cache-4e9d75e8-ca33-4e33-baa3-cb7d596f584e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.537242] env[61868]: DEBUG nova.network.neutron [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 919.567789] env[61868]: DEBUG oslo_concurrency.lockutils [None req-fc9e1fb6-3a19-4eb7-b4e6-935739b166e1 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "interface-5aec2609-79d3-4725-a182-70b575adbe17-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.107s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.622259] env[61868]: DEBUG oslo_vmware.api [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315669, 'name': Rename_Task, 'duration_secs': 0.202785} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.622494] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-127e1772-3b0f-4bda-818a-77c01d320c6e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.624579] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 919.624842] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2cf7305a-d62e-4324-8d83-59164668a77d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.634250] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a72eb7c-eba7-4ecd-8a12-ac704fa941a1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.648204] env[61868]: DEBUG oslo_vmware.api [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 919.648204] env[61868]: value = "task-1315671" [ 919.648204] env[61868]: _type = "Task" [ 919.648204] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.657079] env[61868]: DEBUG oslo_vmware.api [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315671, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.666665] env[61868]: DEBUG nova.compute.manager [req-5f444c96-a62d-4762-b643-e207e84fd349 req-6eee7328-914b-436a-b6ea-0ddbb25eee79 service nova] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Detach interface failed, port_id=4186aca7-63b5-4b24-ab7c-de8f87064242, reason: Instance 306f2dc4-933c-4948-8330-eda93931c4e0 could not be found. {{(pid=61868) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 919.674987] env[61868]: DEBUG oslo_vmware.api [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': task-1315670, 'name': ReconfigVM_Task, 'duration_secs': 0.299101} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.675266] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Reconfigured VM instance instance-00000050 to attach disk [datastore2] b51e968e-6911-47f3-b5a9-e8a5865e42a0/b51e968e-6911-47f3-b5a9-e8a5865e42a0.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 919.675852] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a8f63c0c-8a60-435b-aef6-08c866681184 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.683600] env[61868]: DEBUG oslo_vmware.api [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Waiting for the task: (returnval){ [ 919.683600] env[61868]: value = "task-1315672" [ 919.683600] env[61868]: _type = "Task" [ 919.683600] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.692656] env[61868]: DEBUG oslo_vmware.api [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': task-1315672, 'name': Rename_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.820623] env[61868]: INFO nova.compute.manager [-] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Took 1.33 seconds to deallocate network for instance. [ 919.882442] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4dcc6a68-7a7e-4a85-8566-794266e676cd tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.908s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.884742] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.177s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.886439] env[61868]: INFO nova.compute.claims [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 919.908966] env[61868]: INFO nova.scheduler.client.report [None req-4dcc6a68-7a7e-4a85-8566-794266e676cd tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Deleted allocations for instance c6108cf2-b597-4ca7-8f57-12ea82cc6c2c [ 920.079241] env[61868]: DEBUG nova.network.neutron [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 920.108091] env[61868]: DEBUG oslo_concurrency.lockutils [None req-98db88ff-3992-4d8f-ae3c-e7d7dd0e06c9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Acquiring lock "409cad1a-946d-4c58-aa57-1c0bf97fe63f" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.108377] env[61868]: DEBUG oslo_concurrency.lockutils [None req-98db88ff-3992-4d8f-ae3c-e7d7dd0e06c9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Lock "409cad1a-946d-4c58-aa57-1c0bf97fe63f" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.163936] env[61868]: DEBUG oslo_vmware.api [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315671, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.194018] env[61868]: DEBUG oslo_vmware.api [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': task-1315672, 'name': Rename_Task, 'duration_secs': 0.137402} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.194449] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 920.194655] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-09075106-1181-495f-ace8-f3a8a0c3d2d1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.201207] env[61868]: DEBUG oslo_vmware.api [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Waiting for the task: (returnval){ [ 920.201207] env[61868]: value = "task-1315673" [ 920.201207] env[61868]: _type = "Task" [ 920.201207] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.209493] env[61868]: DEBUG oslo_vmware.api [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': task-1315673, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.281852] env[61868]: DEBUG nova.network.neutron [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Updating instance_info_cache with network_info: [{"id": "ec859835-c58b-4c26-ac77-7628f73990b0", "address": "fa:16:3e:27:fd:8c", "network": {"id": "5b4c5906-b873-4369-855b-e0898ad84812", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-254512637-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "812e115cc1114d0e8536924c4f5556f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec859835-c5", "ovs_interfaceid": "ec859835-c58b-4c26-ac77-7628f73990b0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.327982] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e5888991-048d-4aa8-b4bd-2d8f6de4700f tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.417161] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4dcc6a68-7a7e-4a85-8566-794266e676cd tempest-SecurityGroupsTestJSON-454663905 tempest-SecurityGroupsTestJSON-454663905-project-member] Lock "c6108cf2-b597-4ca7-8f57-12ea82cc6c2c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.925s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.612960] env[61868]: INFO nova.compute.manager [None req-98db88ff-3992-4d8f-ae3c-e7d7dd0e06c9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Detaching volume 7f7cfe03-d391-4d1c-8979-7fb54c701162 [ 920.660123] env[61868]: INFO nova.virt.block_device [None req-98db88ff-3992-4d8f-ae3c-e7d7dd0e06c9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Attempting to driver detach volume 7f7cfe03-d391-4d1c-8979-7fb54c701162 from mountpoint /dev/sdb [ 920.660123] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-98db88ff-3992-4d8f-ae3c-e7d7dd0e06c9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Volume detach. Driver type: vmdk {{(pid=61868) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 920.660123] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-98db88ff-3992-4d8f-ae3c-e7d7dd0e06c9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281562', 'volume_id': '7f7cfe03-d391-4d1c-8979-7fb54c701162', 'name': 'volume-7f7cfe03-d391-4d1c-8979-7fb54c701162', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '409cad1a-946d-4c58-aa57-1c0bf97fe63f', 'attached_at': '', 'detached_at': '', 'volume_id': '7f7cfe03-d391-4d1c-8979-7fb54c701162', 'serial': '7f7cfe03-d391-4d1c-8979-7fb54c701162'} {{(pid=61868) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 920.660123] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e50d3149-2318-4ebf-be7e-6215e2e62ff8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.665187] env[61868]: DEBUG oslo_vmware.api [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315671, 'name': PowerOnVM_Task, 'duration_secs': 1.018131} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.665943] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 920.666191] env[61868]: INFO nova.compute.manager [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Took 8.36 seconds to spawn the instance on the hypervisor. [ 920.666445] env[61868]: DEBUG nova.compute.manager [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 920.667189] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a54e0bd9-e186-4e43-b25d-e6fa38bc3387 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.689521] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ecef54a-3d85-404d-b9bf-37441bcf3fc4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.702197] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b9828b8-8a15-49cc-93f7-18b2734d72da {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.713459] env[61868]: DEBUG oslo_vmware.api [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': task-1315673, 'name': PowerOnVM_Task, 'duration_secs': 0.436497} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.731440] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 920.731746] env[61868]: INFO nova.compute.manager [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Took 5.96 seconds to spawn the instance on the hypervisor. [ 920.732021] env[61868]: DEBUG nova.compute.manager [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 920.732956] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ce305c1-834a-4f6a-9284-0d06dad81846 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.736115] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1010523-3b17-40be-96f8-82c87e103c43 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.759637] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-98db88ff-3992-4d8f-ae3c-e7d7dd0e06c9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] The volume has not been displaced from its original location: [datastore2] volume-7f7cfe03-d391-4d1c-8979-7fb54c701162/volume-7f7cfe03-d391-4d1c-8979-7fb54c701162.vmdk. No consolidation needed. {{(pid=61868) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 920.769022] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-98db88ff-3992-4d8f-ae3c-e7d7dd0e06c9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Reconfiguring VM instance instance-0000003f to detach disk 2001 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 920.769022] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3b0f9ec2-e01a-4554-9106-15d95dbd9728 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.784143] env[61868]: DEBUG oslo_concurrency.lockutils [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Releasing lock "refresh_cache-4e9d75e8-ca33-4e33-baa3-cb7d596f584e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.784450] env[61868]: DEBUG nova.compute.manager [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Instance network_info: |[{"id": "ec859835-c58b-4c26-ac77-7628f73990b0", "address": "fa:16:3e:27:fd:8c", "network": {"id": "5b4c5906-b873-4369-855b-e0898ad84812", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-254512637-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "812e115cc1114d0e8536924c4f5556f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec859835-c5", "ovs_interfaceid": "ec859835-c58b-4c26-ac77-7628f73990b0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 920.784851] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:27:fd:8c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a92a4ffe-7939-4697-bf98-5b22e2c7feda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ec859835-c58b-4c26-ac77-7628f73990b0', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 920.792099] env[61868]: DEBUG oslo.service.loopingcall [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 920.797020] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 920.797020] env[61868]: DEBUG oslo_vmware.api [None req-98db88ff-3992-4d8f-ae3c-e7d7dd0e06c9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Waiting for the task: (returnval){ [ 920.797020] env[61868]: value = "task-1315674" [ 920.797020] env[61868]: _type = "Task" [ 920.797020] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.797020] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-98087c14-8eb5-45aa-afc6-9b530613cc71 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.823613] env[61868]: DEBUG oslo_vmware.api [None req-98db88ff-3992-4d8f-ae3c-e7d7dd0e06c9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315674, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.824960] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 920.824960] env[61868]: value = "task-1315675" [ 920.824960] env[61868]: _type = "Task" [ 920.824960] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.833502] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315675, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.168460] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aef582d5-523b-467b-97f3-5533f16123ba {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.177343] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-561f06f3-b7bf-47e9-82f3-bb3ec9634024 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.217138] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7057a1d6-00d5-46d9-9f22-b38abb137cf6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.224932] env[61868]: INFO nova.compute.manager [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Took 40.71 seconds to build instance. [ 921.231412] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a3cccde-3515-4360-b05e-f3f9eb5bce76 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.248026] env[61868]: DEBUG nova.compute.provider_tree [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 921.276161] env[61868]: INFO nova.compute.manager [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Took 39.18 seconds to build instance. [ 921.284076] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a29eb12d-21de-449b-99db-b45133966326 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Acquiring lock "8ea8e28b-3582-45f2-b4b5-84f624415a58" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.284076] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a29eb12d-21de-449b-99db-b45133966326 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Lock "8ea8e28b-3582-45f2-b4b5-84f624415a58" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.284490] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a29eb12d-21de-449b-99db-b45133966326 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Acquiring lock "8ea8e28b-3582-45f2-b4b5-84f624415a58-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.284490] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a29eb12d-21de-449b-99db-b45133966326 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Lock "8ea8e28b-3582-45f2-b4b5-84f624415a58-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.284590] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a29eb12d-21de-449b-99db-b45133966326 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Lock "8ea8e28b-3582-45f2-b4b5-84f624415a58-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.286624] env[61868]: INFO nova.compute.manager [None req-a29eb12d-21de-449b-99db-b45133966326 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Terminating instance [ 921.290204] env[61868]: DEBUG nova.compute.manager [None req-a29eb12d-21de-449b-99db-b45133966326 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 921.290204] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a29eb12d-21de-449b-99db-b45133966326 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 921.291020] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e45c3842-dc13-45cf-9a4c-709b75d63f28 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.303239] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-a29eb12d-21de-449b-99db-b45133966326 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 921.303529] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-26db827d-10ce-4bfe-86e0-a75945881d57 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.312673] env[61868]: DEBUG oslo_vmware.api [None req-a29eb12d-21de-449b-99db-b45133966326 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 921.312673] env[61868]: value = "task-1315676" [ 921.312673] env[61868]: _type = "Task" [ 921.312673] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.328125] env[61868]: DEBUG oslo_vmware.api [None req-98db88ff-3992-4d8f-ae3c-e7d7dd0e06c9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315674, 'name': ReconfigVM_Task, 'duration_secs': 0.359371} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.335789] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-98db88ff-3992-4d8f-ae3c-e7d7dd0e06c9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Reconfigured VM instance instance-0000003f to detach disk 2001 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 921.340618] env[61868]: DEBUG oslo_vmware.api [None req-a29eb12d-21de-449b-99db-b45133966326 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315676, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.341262] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-077cc57c-b682-459d-a592-d16ae74e7342 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.358276] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315675, 'name': CreateVM_Task, 'duration_secs': 0.363349} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.363022] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 921.363022] env[61868]: DEBUG oslo_vmware.api [None req-98db88ff-3992-4d8f-ae3c-e7d7dd0e06c9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Waiting for the task: (returnval){ [ 921.363022] env[61868]: value = "task-1315677" [ 921.363022] env[61868]: _type = "Task" [ 921.363022] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.363022] env[61868]: DEBUG oslo_concurrency.lockutils [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.363022] env[61868]: DEBUG oslo_concurrency.lockutils [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.363022] env[61868]: DEBUG oslo_concurrency.lockutils [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 921.363479] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-25abcb92-b762-45f8-a326-95c9836d2a3b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.373160] env[61868]: DEBUG oslo_vmware.api [None req-98db88ff-3992-4d8f-ae3c-e7d7dd0e06c9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315677, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.374923] env[61868]: DEBUG oslo_vmware.api [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 921.374923] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]526fe002-edd3-1034-33d6-601e30b021ab" [ 921.374923] env[61868]: _type = "Task" [ 921.374923] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.383671] env[61868]: DEBUG oslo_vmware.api [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]526fe002-edd3-1034-33d6-601e30b021ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.453848] env[61868]: DEBUG nova.compute.manager [req-305320c0-a7a4-4e21-b2e5-5fe809df6b43 req-bc2b4459-7ea9-46f0-af5c-162413e40067 service nova] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Received event network-changed-ec859835-c58b-4c26-ac77-7628f73990b0 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 921.454126] env[61868]: DEBUG nova.compute.manager [req-305320c0-a7a4-4e21-b2e5-5fe809df6b43 req-bc2b4459-7ea9-46f0-af5c-162413e40067 service nova] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Refreshing instance network info cache due to event network-changed-ec859835-c58b-4c26-ac77-7628f73990b0. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 921.454801] env[61868]: DEBUG oslo_concurrency.lockutils [req-305320c0-a7a4-4e21-b2e5-5fe809df6b43 req-bc2b4459-7ea9-46f0-af5c-162413e40067 service nova] Acquiring lock "refresh_cache-4e9d75e8-ca33-4e33-baa3-cb7d596f584e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.454801] env[61868]: DEBUG oslo_concurrency.lockutils [req-305320c0-a7a4-4e21-b2e5-5fe809df6b43 req-bc2b4459-7ea9-46f0-af5c-162413e40067 service nova] Acquired lock "refresh_cache-4e9d75e8-ca33-4e33-baa3-cb7d596f584e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.454801] env[61868]: DEBUG nova.network.neutron [req-305320c0-a7a4-4e21-b2e5-5fe809df6b43 req-bc2b4459-7ea9-46f0-af5c-162413e40067 service nova] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Refreshing network info cache for port ec859835-c58b-4c26-ac77-7628f73990b0 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 921.726805] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c2d65a00-59d1-425c-baa0-29b0a679270f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "e59a182d-97b2-454a-bc40-8afb0839324f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.453s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.750055] env[61868]: DEBUG nova.scheduler.client.report [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 921.777562] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eabc7816-e21a-494b-b050-ec597e57526d tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Lock "b51e968e-6911-47f3-b5a9-e8a5865e42a0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.162s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.788591] env[61868]: DEBUG oslo_concurrency.lockutils [None req-34b56ab0-9afa-4154-aade-bc1aa500c254 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "interface-5aec2609-79d3-4725-a182-70b575adbe17-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.788791] env[61868]: DEBUG oslo_concurrency.lockutils [None req-34b56ab0-9afa-4154-aade-bc1aa500c254 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "interface-5aec2609-79d3-4725-a182-70b575adbe17-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.789176] env[61868]: DEBUG nova.objects.instance [None req-34b56ab0-9afa-4154-aade-bc1aa500c254 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lazy-loading 'flavor' on Instance uuid 5aec2609-79d3-4725-a182-70b575adbe17 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 921.823653] env[61868]: DEBUG oslo_vmware.api [None req-a29eb12d-21de-449b-99db-b45133966326 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315676, 'name': PowerOffVM_Task, 'duration_secs': 0.207668} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.823990] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-a29eb12d-21de-449b-99db-b45133966326 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 921.824717] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a29eb12d-21de-449b-99db-b45133966326 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 921.824986] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3f2858bd-6583-49b3-91fb-e9bd03133150 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.870743] env[61868]: DEBUG oslo_vmware.api [None req-98db88ff-3992-4d8f-ae3c-e7d7dd0e06c9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315677, 'name': ReconfigVM_Task, 'duration_secs': 0.17209} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.871204] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-98db88ff-3992-4d8f-ae3c-e7d7dd0e06c9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281562', 'volume_id': '7f7cfe03-d391-4d1c-8979-7fb54c701162', 'name': 'volume-7f7cfe03-d391-4d1c-8979-7fb54c701162', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '409cad1a-946d-4c58-aa57-1c0bf97fe63f', 'attached_at': '', 'detached_at': '', 'volume_id': '7f7cfe03-d391-4d1c-8979-7fb54c701162', 'serial': '7f7cfe03-d391-4d1c-8979-7fb54c701162'} {{(pid=61868) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 921.891094] env[61868]: DEBUG oslo_vmware.api [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]526fe002-edd3-1034-33d6-601e30b021ab, 'name': SearchDatastore_Task, 'duration_secs': 0.013104} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.891531] env[61868]: DEBUG oslo_concurrency.lockutils [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.891776] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 921.892141] env[61868]: DEBUG oslo_concurrency.lockutils [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.892351] env[61868]: DEBUG oslo_concurrency.lockutils [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.892737] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 921.893147] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4cf379e7-5b4b-4035-818f-a913451a4959 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.903644] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a29eb12d-21de-449b-99db-b45133966326 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 921.903644] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a29eb12d-21de-449b-99db-b45133966326 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 921.903644] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-a29eb12d-21de-449b-99db-b45133966326 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Deleting the datastore file [datastore1] 8ea8e28b-3582-45f2-b4b5-84f624415a58 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 921.903997] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2009791a-5569-4aa0-a06b-cf66d504090e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.907203] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 921.907429] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 921.909323] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f63b30f4-871d-4a06-852c-57613ba904f3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.914911] env[61868]: DEBUG oslo_vmware.api [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 921.914911] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52927376-a232-6d85-ee17-22435be1dd47" [ 921.914911] env[61868]: _type = "Task" [ 921.914911] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.920102] env[61868]: DEBUG oslo_vmware.api [None req-a29eb12d-21de-449b-99db-b45133966326 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 921.920102] env[61868]: value = "task-1315679" [ 921.920102] env[61868]: _type = "Task" [ 921.920102] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.930189] env[61868]: DEBUG oslo_vmware.api [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52927376-a232-6d85-ee17-22435be1dd47, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.936465] env[61868]: DEBUG oslo_vmware.api [None req-a29eb12d-21de-449b-99db-b45133966326 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315679, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.255078] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.370s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.255623] env[61868]: DEBUG nova.compute.manager [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 922.258630] env[61868]: DEBUG oslo_concurrency.lockutils [None req-805d5425-8b50-412b-81b4-c3b60c28ab13 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.825s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.258932] env[61868]: DEBUG nova.objects.instance [None req-805d5425-8b50-412b-81b4-c3b60c28ab13 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lazy-loading 'resources' on Instance uuid b7055e5a-37d2-42d5-bab0-1e70faf52b58 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 922.378044] env[61868]: DEBUG nova.objects.instance [None req-34b56ab0-9afa-4154-aade-bc1aa500c254 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lazy-loading 'pci_requests' on Instance uuid 5aec2609-79d3-4725-a182-70b575adbe17 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 922.407882] env[61868]: DEBUG nova.network.neutron [req-305320c0-a7a4-4e21-b2e5-5fe809df6b43 req-bc2b4459-7ea9-46f0-af5c-162413e40067 service nova] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Updated VIF entry in instance network info cache for port ec859835-c58b-4c26-ac77-7628f73990b0. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 922.408366] env[61868]: DEBUG nova.network.neutron [req-305320c0-a7a4-4e21-b2e5-5fe809df6b43 req-bc2b4459-7ea9-46f0-af5c-162413e40067 service nova] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Updating instance_info_cache with network_info: [{"id": "ec859835-c58b-4c26-ac77-7628f73990b0", "address": "fa:16:3e:27:fd:8c", "network": {"id": "5b4c5906-b873-4369-855b-e0898ad84812", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-254512637-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "812e115cc1114d0e8536924c4f5556f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec859835-c5", "ovs_interfaceid": "ec859835-c58b-4c26-ac77-7628f73990b0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.421300] env[61868]: DEBUG nova.objects.instance [None req-98db88ff-3992-4d8f-ae3c-e7d7dd0e06c9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Lazy-loading 'flavor' on Instance uuid 409cad1a-946d-4c58-aa57-1c0bf97fe63f {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 922.436087] env[61868]: DEBUG oslo_vmware.api [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52927376-a232-6d85-ee17-22435be1dd47, 'name': SearchDatastore_Task, 'duration_secs': 0.014115} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.437441] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-515d0a0d-b6f1-48c6-a679-6cbf7c89bbd0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.443862] env[61868]: DEBUG oslo_vmware.api [None req-a29eb12d-21de-449b-99db-b45133966326 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315679, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.414048} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.444621] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-a29eb12d-21de-449b-99db-b45133966326 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 922.444709] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a29eb12d-21de-449b-99db-b45133966326 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 922.444880] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a29eb12d-21de-449b-99db-b45133966326 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 922.445116] env[61868]: INFO nova.compute.manager [None req-a29eb12d-21de-449b-99db-b45133966326 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Took 1.15 seconds to destroy the instance on the hypervisor. [ 922.445432] env[61868]: DEBUG oslo.service.loopingcall [None req-a29eb12d-21de-449b-99db-b45133966326 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 922.445884] env[61868]: DEBUG nova.compute.manager [-] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 922.446016] env[61868]: DEBUG nova.network.neutron [-] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 922.449405] env[61868]: DEBUG oslo_vmware.api [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 922.449405] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]520a1eb7-dbc4-be0c-9ecb-9a2ee9245eed" [ 922.449405] env[61868]: _type = "Task" [ 922.449405] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.458784] env[61868]: DEBUG oslo_vmware.api [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]520a1eb7-dbc4-be0c-9ecb-9a2ee9245eed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.762673] env[61868]: DEBUG nova.compute.utils [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 922.767258] env[61868]: DEBUG nova.compute.manager [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 922.767694] env[61868]: DEBUG nova.network.neutron [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 922.809508] env[61868]: DEBUG nova.policy [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2f47df2219f64df59cb81d07dd2e79fb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c681bd2f96014872a94455b844546860', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 922.882962] env[61868]: DEBUG nova.objects.base [None req-34b56ab0-9afa-4154-aade-bc1aa500c254 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Object Instance<5aec2609-79d3-4725-a182-70b575adbe17> lazy-loaded attributes: flavor,pci_requests {{(pid=61868) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 922.882962] env[61868]: DEBUG nova.network.neutron [None req-34b56ab0-9afa-4154-aade-bc1aa500c254 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 922.915489] env[61868]: DEBUG oslo_concurrency.lockutils [req-305320c0-a7a4-4e21-b2e5-5fe809df6b43 req-bc2b4459-7ea9-46f0-af5c-162413e40067 service nova] Releasing lock "refresh_cache-4e9d75e8-ca33-4e33-baa3-cb7d596f584e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.962666] env[61868]: DEBUG oslo_vmware.api [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]520a1eb7-dbc4-be0c-9ecb-9a2ee9245eed, 'name': SearchDatastore_Task, 'duration_secs': 0.011811} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.963029] env[61868]: DEBUG nova.policy [None req-34b56ab0-9afa-4154-aade-bc1aa500c254 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c14b4ff01717495ca97c7f35f91c2995', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd7fc76299baf4a97b57139e5f1caa16e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 922.964606] env[61868]: DEBUG oslo_concurrency.lockutils [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.964814] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 4e9d75e8-ca33-4e33-baa3-cb7d596f584e/4e9d75e8-ca33-4e33-baa3-cb7d596f584e.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 922.967782] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-25eeffd7-4308-41f4-a2f3-de106f5dd8ea {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.976695] env[61868]: DEBUG oslo_vmware.api [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 922.976695] env[61868]: value = "task-1315680" [ 922.976695] env[61868]: _type = "Task" [ 922.976695] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.986277] env[61868]: DEBUG oslo_vmware.api [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315680, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.091213] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc36ed90-c3cb-4584-8713-2b9cc4109a23 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.099731] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c478f600-252b-483d-81f5-e1c4e20d82e0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.135424] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ff568a4-f41f-4a0c-ab3a-a5341f2818c1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.141415] env[61868]: DEBUG nova.compute.manager [req-0ade6843-cba9-4534-a550-622e4858d1a6 req-af2fac63-0617-4b19-b77e-80c3241dd673 service nova] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Received event network-changed-cb19d64d-19c9-4d48-8d70-00b3440d9127 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 923.141415] env[61868]: DEBUG nova.compute.manager [req-0ade6843-cba9-4534-a550-622e4858d1a6 req-af2fac63-0617-4b19-b77e-80c3241dd673 service nova] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Refreshing instance network info cache due to event network-changed-cb19d64d-19c9-4d48-8d70-00b3440d9127. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 923.141415] env[61868]: DEBUG oslo_concurrency.lockutils [req-0ade6843-cba9-4534-a550-622e4858d1a6 req-af2fac63-0617-4b19-b77e-80c3241dd673 service nova] Acquiring lock "refresh_cache-e59a182d-97b2-454a-bc40-8afb0839324f" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.141704] env[61868]: DEBUG oslo_concurrency.lockutils [req-0ade6843-cba9-4534-a550-622e4858d1a6 req-af2fac63-0617-4b19-b77e-80c3241dd673 service nova] Acquired lock "refresh_cache-e59a182d-97b2-454a-bc40-8afb0839324f" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.141704] env[61868]: DEBUG nova.network.neutron [req-0ade6843-cba9-4534-a550-622e4858d1a6 req-af2fac63-0617-4b19-b77e-80c3241dd673 service nova] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Refreshing network info cache for port cb19d64d-19c9-4d48-8d70-00b3440d9127 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 923.150794] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09fd5548-81cf-47dd-a0fb-40a9afd447be {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.169308] env[61868]: DEBUG nova.compute.provider_tree [None req-805d5425-8b50-412b-81b4-c3b60c28ab13 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 923.269066] env[61868]: DEBUG nova.compute.manager [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 923.273969] env[61868]: DEBUG nova.network.neutron [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Successfully created port: 61f04f9e-9deb-496a-9301-5db736ced099 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 923.403481] env[61868]: DEBUG nova.network.neutron [-] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.430156] env[61868]: DEBUG oslo_concurrency.lockutils [None req-98db88ff-3992-4d8f-ae3c-e7d7dd0e06c9 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Lock "409cad1a-946d-4c58-aa57-1c0bf97fe63f" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.321s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.489792] env[61868]: DEBUG oslo_vmware.api [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315680, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.509476} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.490356] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 4e9d75e8-ca33-4e33-baa3-cb7d596f584e/4e9d75e8-ca33-4e33-baa3-cb7d596f584e.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 923.490521] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 923.490760] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b962e057-bde1-4ec8-85f5-fbb73cd34e11 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.497805] env[61868]: DEBUG nova.network.neutron [None req-34b56ab0-9afa-4154-aade-bc1aa500c254 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Successfully created port: 8a5f760c-c14e-4701-bde0-97fff3b6289e {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 923.503021] env[61868]: DEBUG oslo_vmware.api [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 923.503021] env[61868]: value = "task-1315681" [ 923.503021] env[61868]: _type = "Task" [ 923.503021] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.510780] env[61868]: DEBUG oslo_vmware.api [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315681, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.660167] env[61868]: INFO nova.compute.manager [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Rebuilding instance [ 923.673542] env[61868]: DEBUG nova.scheduler.client.report [None req-805d5425-8b50-412b-81b4-c3b60c28ab13 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 923.721407] env[61868]: DEBUG nova.compute.manager [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 923.723674] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6b6b39e-7271-4595-ad4d-5a688d868ad9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.906304] env[61868]: INFO nova.compute.manager [-] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Took 1.46 seconds to deallocate network for instance. [ 924.012945] env[61868]: DEBUG oslo_vmware.api [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315681, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07251} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.013226] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 924.014015] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-891e7198-6853-4a4b-a77e-da0c4fefe19d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.038889] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] 4e9d75e8-ca33-4e33-baa3-cb7d596f584e/4e9d75e8-ca33-4e33-baa3-cb7d596f584e.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 924.039265] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4a1ceb50-75cd-4d8c-9319-c624a2d35f49 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.063622] env[61868]: DEBUG oslo_vmware.api [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 924.063622] env[61868]: value = "task-1315682" [ 924.063622] env[61868]: _type = "Task" [ 924.063622] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.073347] env[61868]: DEBUG oslo_vmware.api [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315682, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.182684] env[61868]: DEBUG oslo_concurrency.lockutils [None req-805d5425-8b50-412b-81b4-c3b60c28ab13 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.924s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.189452] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 25.965s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.194655] env[61868]: DEBUG nova.network.neutron [req-0ade6843-cba9-4534-a550-622e4858d1a6 req-af2fac63-0617-4b19-b77e-80c3241dd673 service nova] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Updated VIF entry in instance network info cache for port cb19d64d-19c9-4d48-8d70-00b3440d9127. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 924.195165] env[61868]: DEBUG nova.network.neutron [req-0ade6843-cba9-4534-a550-622e4858d1a6 req-af2fac63-0617-4b19-b77e-80c3241dd673 service nova] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Updating instance_info_cache with network_info: [{"id": "cb19d64d-19c9-4d48-8d70-00b3440d9127", "address": "fa:16:3e:d4:cc:5c", "network": {"id": "3535e048-fe57-4cbc-9489-0b77484069bd", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-748518849-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c8440ce785c344d793dd513b5f008d65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb19d64d-19", "ovs_interfaceid": "cb19d64d-19c9-4d48-8d70-00b3440d9127", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.214454] env[61868]: INFO nova.scheduler.client.report [None req-805d5425-8b50-412b-81b4-c3b60c28ab13 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Deleted allocations for instance b7055e5a-37d2-42d5-bab0-1e70faf52b58 [ 924.235263] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 924.236997] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4b024bc9-ead7-4b28-b2cf-1c97895bba6d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.251499] env[61868]: DEBUG oslo_vmware.api [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Waiting for the task: (returnval){ [ 924.251499] env[61868]: value = "task-1315683" [ 924.251499] env[61868]: _type = "Task" [ 924.251499] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.259937] env[61868]: DEBUG nova.compute.manager [req-656a30c6-ac84-434d-89dd-2abde07e69f8 req-1589436a-014d-4099-b0fc-a5158418a7f1 service nova] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Received event network-vif-deleted-c27d3e06-cf3c-4ef8-ba0b-e6f2ca824646 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 924.264511] env[61868]: DEBUG oslo_vmware.api [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': task-1315683, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.283205] env[61868]: DEBUG nova.compute.manager [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 924.313013] env[61868]: DEBUG nova.virt.hardware [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 924.313372] env[61868]: DEBUG nova.virt.hardware [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 924.313588] env[61868]: DEBUG nova.virt.hardware [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 924.313996] env[61868]: DEBUG nova.virt.hardware [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 924.314102] env[61868]: DEBUG nova.virt.hardware [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 924.314256] env[61868]: DEBUG nova.virt.hardware [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 924.314552] env[61868]: DEBUG nova.virt.hardware [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 924.314781] env[61868]: DEBUG nova.virt.hardware [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 924.314988] env[61868]: DEBUG nova.virt.hardware [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 924.315252] env[61868]: DEBUG nova.virt.hardware [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 924.315506] env[61868]: DEBUG nova.virt.hardware [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 924.316963] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-753f6b73-948c-4c7a-b019-5023b82c0959 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.328018] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bee6d108-0cce-4213-95af-82471d66aaa9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.416360] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a29eb12d-21de-449b-99db-b45133966326 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.575921] env[61868]: DEBUG oslo_vmware.api [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315682, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.698233] env[61868]: DEBUG oslo_concurrency.lockutils [req-0ade6843-cba9-4534-a550-622e4858d1a6 req-af2fac63-0617-4b19-b77e-80c3241dd673 service nova] Releasing lock "refresh_cache-e59a182d-97b2-454a-bc40-8afb0839324f" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.726826] env[61868]: DEBUG oslo_concurrency.lockutils [None req-805d5425-8b50-412b-81b4-c3b60c28ab13 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "b7055e5a-37d2-42d5-bab0-1e70faf52b58" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.397s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.744941] env[61868]: DEBUG oslo_concurrency.lockutils [None req-81536fac-7f2f-4c97-955a-3186c67a3a47 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Acquiring lock "409cad1a-946d-4c58-aa57-1c0bf97fe63f" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.745216] env[61868]: DEBUG oslo_concurrency.lockutils [None req-81536fac-7f2f-4c97-955a-3186c67a3a47 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Lock "409cad1a-946d-4c58-aa57-1c0bf97fe63f" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.762367] env[61868]: DEBUG oslo_vmware.api [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': task-1315683, 'name': PowerOffVM_Task, 'duration_secs': 0.179326} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.763039] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 924.763039] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 924.763681] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36f31d08-216d-4785-98ba-4c3b6e302e91 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.771325] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 924.771408] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9c122e9e-538f-4ee9-b422-6e50f4793671 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.798292] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 924.798612] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Deleting contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 924.798842] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Deleting the datastore file [datastore2] b51e968e-6911-47f3-b5a9-e8a5865e42a0 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 924.799174] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-00a0a1f2-7ca5-497a-bfc8-e5bc032b264a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.806943] env[61868]: DEBUG oslo_vmware.api [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Waiting for the task: (returnval){ [ 924.806943] env[61868]: value = "task-1315685" [ 924.806943] env[61868]: _type = "Task" [ 924.806943] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.816534] env[61868]: DEBUG oslo_vmware.api [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': task-1315685, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.075453] env[61868]: DEBUG oslo_vmware.api [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315682, 'name': ReconfigVM_Task, 'duration_secs': 0.905743} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.075531] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Reconfigured VM instance instance-00000051 to attach disk [datastore1] 4e9d75e8-ca33-4e33-baa3-cb7d596f584e/4e9d75e8-ca33-4e33-baa3-cb7d596f584e.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 925.076159] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a95bc78b-d483-4511-960c-83f0ee17300f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.084810] env[61868]: DEBUG oslo_vmware.api [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 925.084810] env[61868]: value = "task-1315686" [ 925.084810] env[61868]: _type = "Task" [ 925.084810] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.095554] env[61868]: DEBUG oslo_vmware.api [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315686, 'name': Rename_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.202117] env[61868]: INFO nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Updating resource usage from migration 70dd0d99-e65a-47fd-872a-4c6b316b0d07 [ 925.230369] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 409cad1a-946d-4c58-aa57-1c0bf97fe63f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 925.230369] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 19c2720c-90bc-47f6-999b-6031f893408d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 925.230369] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 75fdbc11-1ef3-442d-bcf4-55069ba9ab62 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 925.230369] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 5aec2609-79d3-4725-a182-70b575adbe17 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 925.230509] env[61868]: WARNING nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 925.230509] env[61868]: WARNING nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 306f2dc4-933c-4948-8330-eda93931c4e0 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 925.230509] env[61868]: WARNING nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance e2832c0f-dc62-4628-87d4-e2a01819b771 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 925.230509] env[61868]: WARNING nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 8ea8e28b-3582-45f2-b4b5-84f624415a58 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 925.230509] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance e59a182d-97b2-454a-bc40-8afb0839324f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 925.230662] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance b51e968e-6911-47f3-b5a9-e8a5865e42a0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 925.230662] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 4e9d75e8-ca33-4e33-baa3-cb7d596f584e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 925.230662] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 925.238452] env[61868]: DEBUG nova.network.neutron [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Successfully updated port: 61f04f9e-9deb-496a-9301-5db736ced099 {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 925.247960] env[61868]: INFO nova.compute.manager [None req-81536fac-7f2f-4c97-955a-3186c67a3a47 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Detaching volume fdcd4a1e-341a-4607-b405-3e3c2a834712 [ 925.288573] env[61868]: DEBUG nova.network.neutron [None req-34b56ab0-9afa-4154-aade-bc1aa500c254 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Successfully updated port: 8a5f760c-c14e-4701-bde0-97fff3b6289e {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 925.296390] env[61868]: INFO nova.virt.block_device [None req-81536fac-7f2f-4c97-955a-3186c67a3a47 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Attempting to driver detach volume fdcd4a1e-341a-4607-b405-3e3c2a834712 from mountpoint /dev/sdc [ 925.296637] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-81536fac-7f2f-4c97-955a-3186c67a3a47 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Volume detach. Driver type: vmdk {{(pid=61868) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 925.296869] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-81536fac-7f2f-4c97-955a-3186c67a3a47 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281570', 'volume_id': 'fdcd4a1e-341a-4607-b405-3e3c2a834712', 'name': 'volume-fdcd4a1e-341a-4607-b405-3e3c2a834712', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '409cad1a-946d-4c58-aa57-1c0bf97fe63f', 'attached_at': '', 'detached_at': '', 'volume_id': 'fdcd4a1e-341a-4607-b405-3e3c2a834712', 'serial': 'fdcd4a1e-341a-4607-b405-3e3c2a834712'} {{(pid=61868) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 925.298057] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-097a22c3-6e1a-44ab-a2c5-a375fc089f44 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.328402] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b8567c1-662e-440d-8466-185d6e3c4383 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.339066] env[61868]: DEBUG oslo_vmware.api [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': task-1315685, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.462821} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.340931] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 925.341180] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Deleted contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 925.341380] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 925.345668] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06712723-8d31-40d3-8e62-bfad4c7c940e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.370621] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c776efd2-a8ef-4118-9309-716733b46316 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.387318] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-81536fac-7f2f-4c97-955a-3186c67a3a47 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] The volume has not been displaced from its original location: [datastore2] volume-fdcd4a1e-341a-4607-b405-3e3c2a834712/volume-fdcd4a1e-341a-4607-b405-3e3c2a834712.vmdk. No consolidation needed. {{(pid=61868) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 925.392635] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-81536fac-7f2f-4c97-955a-3186c67a3a47 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Reconfiguring VM instance instance-0000003f to detach disk 2002 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 925.393847] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e20260f7-18b9-48c4-a1c6-ef307f412744 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.415193] env[61868]: DEBUG oslo_vmware.api [None req-81536fac-7f2f-4c97-955a-3186c67a3a47 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Waiting for the task: (returnval){ [ 925.415193] env[61868]: value = "task-1315687" [ 925.415193] env[61868]: _type = "Task" [ 925.415193] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.428727] env[61868]: DEBUG oslo_vmware.api [None req-81536fac-7f2f-4c97-955a-3186c67a3a47 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315687, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.591941] env[61868]: DEBUG nova.compute.manager [req-5ad29ae3-10e7-4232-a17b-fd5b6c601af4 req-61792e42-5517-4de0-b607-755447c67fc7 service nova] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Received event network-vif-plugged-61f04f9e-9deb-496a-9301-5db736ced099 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 925.592218] env[61868]: DEBUG oslo_concurrency.lockutils [req-5ad29ae3-10e7-4232-a17b-fd5b6c601af4 req-61792e42-5517-4de0-b607-755447c67fc7 service nova] Acquiring lock "4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.592640] env[61868]: DEBUG oslo_concurrency.lockutils [req-5ad29ae3-10e7-4232-a17b-fd5b6c601af4 req-61792e42-5517-4de0-b607-755447c67fc7 service nova] Lock "4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.592640] env[61868]: DEBUG oslo_concurrency.lockutils [req-5ad29ae3-10e7-4232-a17b-fd5b6c601af4 req-61792e42-5517-4de0-b607-755447c67fc7 service nova] Lock "4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.592809] env[61868]: DEBUG nova.compute.manager [req-5ad29ae3-10e7-4232-a17b-fd5b6c601af4 req-61792e42-5517-4de0-b607-755447c67fc7 service nova] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] No waiting events found dispatching network-vif-plugged-61f04f9e-9deb-496a-9301-5db736ced099 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 925.592921] env[61868]: WARNING nova.compute.manager [req-5ad29ae3-10e7-4232-a17b-fd5b6c601af4 req-61792e42-5517-4de0-b607-755447c67fc7 service nova] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Received unexpected event network-vif-plugged-61f04f9e-9deb-496a-9301-5db736ced099 for instance with vm_state building and task_state spawning. [ 925.593160] env[61868]: DEBUG nova.compute.manager [req-5ad29ae3-10e7-4232-a17b-fd5b6c601af4 req-61792e42-5517-4de0-b607-755447c67fc7 service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Received event network-vif-plugged-8a5f760c-c14e-4701-bde0-97fff3b6289e {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 925.593342] env[61868]: DEBUG oslo_concurrency.lockutils [req-5ad29ae3-10e7-4232-a17b-fd5b6c601af4 req-61792e42-5517-4de0-b607-755447c67fc7 service nova] Acquiring lock "5aec2609-79d3-4725-a182-70b575adbe17-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.593528] env[61868]: DEBUG oslo_concurrency.lockutils [req-5ad29ae3-10e7-4232-a17b-fd5b6c601af4 req-61792e42-5517-4de0-b607-755447c67fc7 service nova] Lock "5aec2609-79d3-4725-a182-70b575adbe17-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.593689] env[61868]: DEBUG oslo_concurrency.lockutils [req-5ad29ae3-10e7-4232-a17b-fd5b6c601af4 req-61792e42-5517-4de0-b607-755447c67fc7 service nova] Lock "5aec2609-79d3-4725-a182-70b575adbe17-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.593941] env[61868]: DEBUG nova.compute.manager [req-5ad29ae3-10e7-4232-a17b-fd5b6c601af4 req-61792e42-5517-4de0-b607-755447c67fc7 service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] No waiting events found dispatching network-vif-plugged-8a5f760c-c14e-4701-bde0-97fff3b6289e {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 925.594026] env[61868]: WARNING nova.compute.manager [req-5ad29ae3-10e7-4232-a17b-fd5b6c601af4 req-61792e42-5517-4de0-b607-755447c67fc7 service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Received unexpected event network-vif-plugged-8a5f760c-c14e-4701-bde0-97fff3b6289e for instance with vm_state active and task_state None. [ 925.600631] env[61868]: DEBUG oslo_vmware.api [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315686, 'name': Rename_Task, 'duration_secs': 0.288872} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.600905] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 925.601185] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-df849790-c1e2-462e-8ed0-42776d14f153 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.608419] env[61868]: DEBUG oslo_vmware.api [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 925.608419] env[61868]: value = "task-1315688" [ 925.608419] env[61868]: _type = "Task" [ 925.608419] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.616929] env[61868]: DEBUG oslo_vmware.api [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315688, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.733035] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 6ce39c07-939c-49f3-8871-6e2163709f4a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 925.733305] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Migration 70dd0d99-e65a-47fd-872a-4c6b316b0d07 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 925.733435] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 842c8606-632a-47d4-9a67-38d8e6ce6b4d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 925.740647] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Acquiring lock "refresh_cache-4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.740934] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Acquired lock "refresh_cache-4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.741113] env[61868]: DEBUG nova.network.neutron [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 925.792082] env[61868]: DEBUG oslo_concurrency.lockutils [None req-34b56ab0-9afa-4154-aade-bc1aa500c254 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "refresh_cache-5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.792498] env[61868]: DEBUG oslo_concurrency.lockutils [None req-34b56ab0-9afa-4154-aade-bc1aa500c254 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquired lock "refresh_cache-5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.792857] env[61868]: DEBUG nova.network.neutron [None req-34b56ab0-9afa-4154-aade-bc1aa500c254 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 925.927155] env[61868]: DEBUG oslo_vmware.api [None req-81536fac-7f2f-4c97-955a-3186c67a3a47 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315687, 'name': ReconfigVM_Task, 'duration_secs': 0.345938} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.927460] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-81536fac-7f2f-4c97-955a-3186c67a3a47 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Reconfigured VM instance instance-0000003f to detach disk 2002 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 925.932369] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0a2ab164-34df-4760-9e40-3d791f569194 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.949599] env[61868]: DEBUG oslo_vmware.api [None req-81536fac-7f2f-4c97-955a-3186c67a3a47 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Waiting for the task: (returnval){ [ 925.949599] env[61868]: value = "task-1315689" [ 925.949599] env[61868]: _type = "Task" [ 925.949599] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.958611] env[61868]: DEBUG oslo_vmware.api [None req-81536fac-7f2f-4c97-955a-3186c67a3a47 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315689, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.067793] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.068046] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.118761] env[61868]: DEBUG oslo_vmware.api [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315688, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.236784] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 87b25494-1853-4c7b-ae8b-6283992e895c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 926.275930] env[61868]: DEBUG nova.network.neutron [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 926.342028] env[61868]: WARNING nova.network.neutron [None req-34b56ab0-9afa-4154-aade-bc1aa500c254 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] d4f100cd-8de6-4459-a7a6-bc1895b8b14f already exists in list: networks containing: ['d4f100cd-8de6-4459-a7a6-bc1895b8b14f']. ignoring it [ 926.382306] env[61868]: DEBUG nova.virt.hardware [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 926.383035] env[61868]: DEBUG nova.virt.hardware [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 926.383035] env[61868]: DEBUG nova.virt.hardware [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 926.383261] env[61868]: DEBUG nova.virt.hardware [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 926.383432] env[61868]: DEBUG nova.virt.hardware [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 926.383574] env[61868]: DEBUG nova.virt.hardware [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 926.383789] env[61868]: DEBUG nova.virt.hardware [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 926.383961] env[61868]: DEBUG nova.virt.hardware [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 926.384155] env[61868]: DEBUG nova.virt.hardware [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 926.384326] env[61868]: DEBUG nova.virt.hardware [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 926.384671] env[61868]: DEBUG nova.virt.hardware [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 926.385736] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c81e6ca-cc78-4171-bc0d-508d568e91e6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.395946] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92bb61ec-3985-4b7c-bb99-14d6add87b85 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.412128] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Instance VIF info [] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 926.417733] env[61868]: DEBUG oslo.service.loopingcall [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 926.422313] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 926.422571] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-74dcb212-d4ef-4cef-9dec-dc5439f080ee {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.440470] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 926.440470] env[61868]: value = "task-1315690" [ 926.440470] env[61868]: _type = "Task" [ 926.440470] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.450171] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315690, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.459145] env[61868]: DEBUG oslo_vmware.api [None req-81536fac-7f2f-4c97-955a-3186c67a3a47 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315689, 'name': ReconfigVM_Task, 'duration_secs': 0.173536} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.459494] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-81536fac-7f2f-4c97-955a-3186c67a3a47 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281570', 'volume_id': 'fdcd4a1e-341a-4607-b405-3e3c2a834712', 'name': 'volume-fdcd4a1e-341a-4607-b405-3e3c2a834712', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '409cad1a-946d-4c58-aa57-1c0bf97fe63f', 'attached_at': '', 'detached_at': '', 'volume_id': 'fdcd4a1e-341a-4607-b405-3e3c2a834712', 'serial': 'fdcd4a1e-341a-4607-b405-3e3c2a834712'} {{(pid=61868) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 926.568263] env[61868]: DEBUG nova.network.neutron [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Updating instance_info_cache with network_info: [{"id": "61f04f9e-9deb-496a-9301-5db736ced099", "address": "fa:16:3e:0d:59:c5", "network": {"id": "2adf7e94-33eb-44fd-a6a4-59b2beb67b5e", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-2079371774-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c681bd2f96014872a94455b844546860", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61f04f9e-9d", "ovs_interfaceid": "61f04f9e-9deb-496a-9301-5db736ced099", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.569666] env[61868]: DEBUG nova.compute.manager [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 926.621259] env[61868]: DEBUG oslo_vmware.api [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315688, 'name': PowerOnVM_Task, 'duration_secs': 0.801051} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.621541] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 926.621745] env[61868]: INFO nova.compute.manager [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Took 8.77 seconds to spawn the instance on the hypervisor. [ 926.621927] env[61868]: DEBUG nova.compute.manager [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 926.622755] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa09b110-cf79-41cc-8e63-64843f295671 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.740148] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 867384c6-175b-4848-b9ac-917470742ba2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 926.740318] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Total usable vcpus: 48, total allocated vcpus: 9 {{(pid=61868) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 926.740466] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2240MB phys_disk=200GB used_disk=9GB total_vcpus=48 used_vcpus=9 pci_stats=[] {{(pid=61868) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 926.760552] env[61868]: DEBUG nova.network.neutron [None req-34b56ab0-9afa-4154-aade-bc1aa500c254 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Updating instance_info_cache with network_info: [{"id": "9771cedf-dae3-41e8-ad4b-aec421274d6c", "address": "fa:16:3e:4f:5f:b8", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.244", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9771cedf-da", "ovs_interfaceid": "9771cedf-dae3-41e8-ad4b-aec421274d6c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "8a5f760c-c14e-4701-bde0-97fff3b6289e", "address": "fa:16:3e:f7:7a:b3", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8a5f760c-c1", "ovs_interfaceid": "8a5f760c-c14e-4701-bde0-97fff3b6289e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.956957] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315690, 'name': CreateVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.011593] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3e96e93-9d3f-4a39-9648-8707aee60f0a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.015872] env[61868]: DEBUG nova.objects.instance [None req-81536fac-7f2f-4c97-955a-3186c67a3a47 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Lazy-loading 'flavor' on Instance uuid 409cad1a-946d-4c58-aa57-1c0bf97fe63f {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 927.024556] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e6fd0cd-009d-45b1-854a-6af6a6c6d325 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.055749] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-431f1498-d36e-48a8-9934-8c3e4e87451a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.065436] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb5a6210-eca9-4bc5-81f7-8f216c69ce1b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.080979] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Releasing lock "refresh_cache-4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.081392] env[61868]: DEBUG nova.compute.manager [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Instance network_info: |[{"id": "61f04f9e-9deb-496a-9301-5db736ced099", "address": "fa:16:3e:0d:59:c5", "network": {"id": "2adf7e94-33eb-44fd-a6a4-59b2beb67b5e", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-2079371774-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c681bd2f96014872a94455b844546860", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61f04f9e-9d", "ovs_interfaceid": "61f04f9e-9deb-496a-9301-5db736ced099", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 927.084604] env[61868]: DEBUG nova.compute.provider_tree [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 927.086377] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0d:59:c5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3d062877-0b23-4965-908b-f585f25f3bf1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '61f04f9e-9deb-496a-9301-5db736ced099', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 927.095309] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Creating folder: Project (c681bd2f96014872a94455b844546860). Parent ref: group-v281478. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 927.099295] env[61868]: DEBUG nova.scheduler.client.report [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 927.103812] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6e7a4793-8a74-4fd6-84b8-3bb3c5addcb1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.122848] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Created folder: Project (c681bd2f96014872a94455b844546860) in parent group-v281478. [ 927.122848] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Creating folder: Instances. Parent ref: group-v281579. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 927.122848] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7fc8db8b-4644-4f0b-9bb0-cf6365608dd4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.124709] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.140582] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Created folder: Instances in parent group-v281579. [ 927.140582] env[61868]: DEBUG oslo.service.loopingcall [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 927.142076] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 927.142652] env[61868]: INFO nova.compute.manager [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Took 42.55 seconds to build instance. [ 927.144076] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7754b95c-cdc3-42bd-b0ca-f286d2d1b401 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.168346] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 927.168346] env[61868]: value = "task-1315693" [ 927.168346] env[61868]: _type = "Task" [ 927.168346] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.184451] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315693, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.263761] env[61868]: DEBUG oslo_concurrency.lockutils [None req-34b56ab0-9afa-4154-aade-bc1aa500c254 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Releasing lock "refresh_cache-5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.264802] env[61868]: DEBUG oslo_concurrency.lockutils [None req-34b56ab0-9afa-4154-aade-bc1aa500c254 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.265114] env[61868]: DEBUG oslo_concurrency.lockutils [None req-34b56ab0-9afa-4154-aade-bc1aa500c254 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquired lock "5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.266829] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd4c45f2-363b-4fe0-81dc-23b40132d305 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.300394] env[61868]: DEBUG nova.virt.hardware [None req-34b56ab0-9afa-4154-aade-bc1aa500c254 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 927.300857] env[61868]: DEBUG nova.virt.hardware [None req-34b56ab0-9afa-4154-aade-bc1aa500c254 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 927.301061] env[61868]: DEBUG nova.virt.hardware [None req-34b56ab0-9afa-4154-aade-bc1aa500c254 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 927.301369] env[61868]: DEBUG nova.virt.hardware [None req-34b56ab0-9afa-4154-aade-bc1aa500c254 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 927.301645] env[61868]: DEBUG nova.virt.hardware [None req-34b56ab0-9afa-4154-aade-bc1aa500c254 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 927.301913] env[61868]: DEBUG nova.virt.hardware [None req-34b56ab0-9afa-4154-aade-bc1aa500c254 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 927.302295] env[61868]: DEBUG nova.virt.hardware [None req-34b56ab0-9afa-4154-aade-bc1aa500c254 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 927.302579] env[61868]: DEBUG nova.virt.hardware [None req-34b56ab0-9afa-4154-aade-bc1aa500c254 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 927.302883] env[61868]: DEBUG nova.virt.hardware [None req-34b56ab0-9afa-4154-aade-bc1aa500c254 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 927.303646] env[61868]: DEBUG nova.virt.hardware [None req-34b56ab0-9afa-4154-aade-bc1aa500c254 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 927.303840] env[61868]: DEBUG nova.virt.hardware [None req-34b56ab0-9afa-4154-aade-bc1aa500c254 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 927.316549] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-34b56ab0-9afa-4154-aade-bc1aa500c254 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Reconfiguring VM to attach interface {{(pid=61868) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 927.317588] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-91c2a543-3ba2-4dc2-8ada-c564cf367bc1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.341019] env[61868]: DEBUG oslo_vmware.api [None req-34b56ab0-9afa-4154-aade-bc1aa500c254 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for the task: (returnval){ [ 927.341019] env[61868]: value = "task-1315694" [ 927.341019] env[61868]: _type = "Task" [ 927.341019] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.350533] env[61868]: DEBUG oslo_vmware.api [None req-34b56ab0-9afa-4154-aade-bc1aa500c254 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315694, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.455848] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315690, 'name': CreateVM_Task, 'duration_secs': 0.94695} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.455848] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 927.455848] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.455848] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.455848] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 927.455848] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a78d0501-4e8d-43bd-931b-0a347a879f26 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.461422] env[61868]: DEBUG oslo_vmware.api [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Waiting for the task: (returnval){ [ 927.461422] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52c7a152-269f-b1b4-6f76-8f0f500f4ef7" [ 927.461422] env[61868]: _type = "Task" [ 927.461422] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.469358] env[61868]: DEBUG oslo_vmware.api [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52c7a152-269f-b1b4-6f76-8f0f500f4ef7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.608579] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61868) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 927.608834] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.423s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.609148] env[61868]: DEBUG oslo_concurrency.lockutils [None req-780bfa57-3788-4c6b-93f4-f83d47616fd0 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.832s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.609352] env[61868]: DEBUG oslo_concurrency.lockutils [None req-780bfa57-3788-4c6b-93f4-f83d47616fd0 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.611516] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.805s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.613546] env[61868]: INFO nova.compute.claims [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 927.647475] env[61868]: INFO nova.scheduler.client.report [None req-780bfa57-3788-4c6b-93f4-f83d47616fd0 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Deleted allocations for instance e2832c0f-dc62-4628-87d4-e2a01819b771 [ 927.649160] env[61868]: DEBUG oslo_concurrency.lockutils [None req-238dac98-3cdf-4125-9866-99be864d092f tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "4e9d75e8-ca33-4e33-baa3-cb7d596f584e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.206s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.682531] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315693, 'name': CreateVM_Task} progress is 25%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.733572] env[61868]: DEBUG nova.compute.manager [req-f1c13f2e-b1e5-43b0-b50d-b2fd18d435ac req-41e5706e-f327-4097-9090-b436db93a2e6 service nova] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Received event network-changed-61f04f9e-9deb-496a-9301-5db736ced099 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 927.733739] env[61868]: DEBUG nova.compute.manager [req-f1c13f2e-b1e5-43b0-b50d-b2fd18d435ac req-41e5706e-f327-4097-9090-b436db93a2e6 service nova] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Refreshing instance network info cache due to event network-changed-61f04f9e-9deb-496a-9301-5db736ced099. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 927.733964] env[61868]: DEBUG oslo_concurrency.lockutils [req-f1c13f2e-b1e5-43b0-b50d-b2fd18d435ac req-41e5706e-f327-4097-9090-b436db93a2e6 service nova] Acquiring lock "refresh_cache-4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.734140] env[61868]: DEBUG oslo_concurrency.lockutils [req-f1c13f2e-b1e5-43b0-b50d-b2fd18d435ac req-41e5706e-f327-4097-9090-b436db93a2e6 service nova] Acquired lock "refresh_cache-4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.734309] env[61868]: DEBUG nova.network.neutron [req-f1c13f2e-b1e5-43b0-b50d-b2fd18d435ac req-41e5706e-f327-4097-9090-b436db93a2e6 service nova] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Refreshing network info cache for port 61f04f9e-9deb-496a-9301-5db736ced099 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 927.852356] env[61868]: DEBUG oslo_vmware.api [None req-34b56ab0-9afa-4154-aade-bc1aa500c254 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315694, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.971549] env[61868]: DEBUG oslo_vmware.api [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52c7a152-269f-b1b4-6f76-8f0f500f4ef7, 'name': SearchDatastore_Task, 'duration_secs': 0.015155} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.972069] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.972448] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 927.972817] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.973229] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.973570] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 927.974088] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-54d43bc4-84e8-4ada-8dac-94a92d86c673 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.988208] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 927.988208] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 927.990275] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b48e1f2b-54c2-42e9-915d-7a883a9237ed {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.997668] env[61868]: DEBUG oslo_vmware.api [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Waiting for the task: (returnval){ [ 927.997668] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e31460-be9d-164d-fab0-3e241474cba2" [ 927.997668] env[61868]: _type = "Task" [ 927.997668] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.010045] env[61868]: DEBUG oslo_vmware.api [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e31460-be9d-164d-fab0-3e241474cba2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.025188] env[61868]: DEBUG oslo_concurrency.lockutils [None req-81536fac-7f2f-4c97-955a-3186c67a3a47 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Lock "409cad1a-946d-4c58-aa57-1c0bf97fe63f" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.280s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.162430] env[61868]: DEBUG oslo_concurrency.lockutils [None req-780bfa57-3788-4c6b-93f4-f83d47616fd0 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lock "e2832c0f-dc62-4628-87d4-e2a01819b771" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.679s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.184434] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315693, 'name': CreateVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.357103] env[61868]: DEBUG oslo_vmware.api [None req-34b56ab0-9afa-4154-aade-bc1aa500c254 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315694, 'name': ReconfigVM_Task, 'duration_secs': 0.897136} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.357103] env[61868]: DEBUG oslo_concurrency.lockutils [None req-34b56ab0-9afa-4154-aade-bc1aa500c254 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Releasing lock "5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.357103] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-34b56ab0-9afa-4154-aade-bc1aa500c254 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Reconfigured VM to attach interface {{(pid=61868) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 928.488093] env[61868]: DEBUG oslo_concurrency.lockutils [None req-40f8b9a6-0376-496a-97a7-807a4e018476 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Acquiring lock "409cad1a-946d-4c58-aa57-1c0bf97fe63f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.488402] env[61868]: DEBUG oslo_concurrency.lockutils [None req-40f8b9a6-0376-496a-97a7-807a4e018476 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Lock "409cad1a-946d-4c58-aa57-1c0bf97fe63f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.488623] env[61868]: DEBUG oslo_concurrency.lockutils [None req-40f8b9a6-0376-496a-97a7-807a4e018476 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Acquiring lock "409cad1a-946d-4c58-aa57-1c0bf97fe63f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.488935] env[61868]: DEBUG oslo_concurrency.lockutils [None req-40f8b9a6-0376-496a-97a7-807a4e018476 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Lock "409cad1a-946d-4c58-aa57-1c0bf97fe63f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.489033] env[61868]: DEBUG oslo_concurrency.lockutils [None req-40f8b9a6-0376-496a-97a7-807a4e018476 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Lock "409cad1a-946d-4c58-aa57-1c0bf97fe63f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.492777] env[61868]: INFO nova.compute.manager [None req-40f8b9a6-0376-496a-97a7-807a4e018476 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Terminating instance [ 928.495126] env[61868]: DEBUG nova.compute.manager [None req-40f8b9a6-0376-496a-97a7-807a4e018476 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 928.495372] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-40f8b9a6-0376-496a-97a7-807a4e018476 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 928.496380] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75392ed9-b001-4ee0-9374-1a050eddb130 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.513853] env[61868]: DEBUG oslo_vmware.api [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e31460-be9d-164d-fab0-3e241474cba2, 'name': SearchDatastore_Task, 'duration_secs': 0.013451} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.514253] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-40f8b9a6-0376-496a-97a7-807a4e018476 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 928.514932] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ff6ed4b5-e5b9-4b47-af9b-341df50adabf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.516482] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-58e8c8fd-84d0-4333-8dab-fb89f8def455 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.523179] env[61868]: DEBUG oslo_vmware.api [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Waiting for the task: (returnval){ [ 928.523179] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52a642b8-e76e-1d16-b4f6-167ec6c5bd3b" [ 928.523179] env[61868]: _type = "Task" [ 928.523179] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.528188] env[61868]: DEBUG oslo_vmware.api [None req-40f8b9a6-0376-496a-97a7-807a4e018476 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Waiting for the task: (returnval){ [ 928.528188] env[61868]: value = "task-1315695" [ 928.528188] env[61868]: _type = "Task" [ 928.528188] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.535246] env[61868]: DEBUG oslo_vmware.api [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52a642b8-e76e-1d16-b4f6-167ec6c5bd3b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.544796] env[61868]: DEBUG oslo_vmware.api [None req-40f8b9a6-0376-496a-97a7-807a4e018476 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315695, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.601061] env[61868]: DEBUG nova.network.neutron [req-f1c13f2e-b1e5-43b0-b50d-b2fd18d435ac req-41e5706e-f327-4097-9090-b436db93a2e6 service nova] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Updated VIF entry in instance network info cache for port 61f04f9e-9deb-496a-9301-5db736ced099. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 928.601534] env[61868]: DEBUG nova.network.neutron [req-f1c13f2e-b1e5-43b0-b50d-b2fd18d435ac req-41e5706e-f327-4097-9090-b436db93a2e6 service nova] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Updating instance_info_cache with network_info: [{"id": "61f04f9e-9deb-496a-9301-5db736ced099", "address": "fa:16:3e:0d:59:c5", "network": {"id": "2adf7e94-33eb-44fd-a6a4-59b2beb67b5e", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-2079371774-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c681bd2f96014872a94455b844546860", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61f04f9e-9d", "ovs_interfaceid": "61f04f9e-9deb-496a-9301-5db736ced099", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.607838] env[61868]: DEBUG oslo_concurrency.lockutils [None req-23a74488-062f-46a6-a299-a91a3d25288d tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "4e9d75e8-ca33-4e33-baa3-cb7d596f584e" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.608073] env[61868]: DEBUG oslo_concurrency.lockutils [None req-23a74488-062f-46a6-a299-a91a3d25288d tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "4e9d75e8-ca33-4e33-baa3-cb7d596f584e" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.608261] env[61868]: DEBUG nova.compute.manager [None req-23a74488-062f-46a6-a299-a91a3d25288d tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 928.609280] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5678225-ca27-42b5-af07-7cdb2d3340a3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.619662] env[61868]: DEBUG nova.compute.manager [None req-23a74488-062f-46a6-a299-a91a3d25288d tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61868) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 928.620275] env[61868]: DEBUG nova.objects.instance [None req-23a74488-062f-46a6-a299-a91a3d25288d tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lazy-loading 'flavor' on Instance uuid 4e9d75e8-ca33-4e33-baa3-cb7d596f584e {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 928.680855] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315693, 'name': CreateVM_Task, 'duration_secs': 1.138143} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.681039] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 928.681753] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.681928] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.682321] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 928.682594] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb66020a-ba9b-4f7e-898a-974cf8e886a9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.688721] env[61868]: DEBUG oslo_vmware.api [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Waiting for the task: (returnval){ [ 928.688721] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52ec8407-be98-f445-736c-5e7cc89bf113" [ 928.688721] env[61868]: _type = "Task" [ 928.688721] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.701401] env[61868]: DEBUG oslo_vmware.api [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52ec8407-be98-f445-736c-5e7cc89bf113, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.864704] env[61868]: DEBUG oslo_concurrency.lockutils [None req-34b56ab0-9afa-4154-aade-bc1aa500c254 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "interface-5aec2609-79d3-4725-a182-70b575adbe17-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.076s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.881082] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e89c1c1-e014-47fb-a7de-e92bca608f26 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.890455] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cf6ae83-f5d0-44bc-a6a9-a9e5811ea9ce {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.925246] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a26cbb0-6cf3-4ad8-9575-53cd2ba3c48e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.933650] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c4692fd-2605-4499-8671-25e97331bd91 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.949030] env[61868]: DEBUG nova.compute.provider_tree [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 929.038489] env[61868]: DEBUG oslo_vmware.api [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52a642b8-e76e-1d16-b4f6-167ec6c5bd3b, 'name': SearchDatastore_Task, 'duration_secs': 0.018881} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.039298] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.039635] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] b51e968e-6911-47f3-b5a9-e8a5865e42a0/b51e968e-6911-47f3-b5a9-e8a5865e42a0.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 929.039959] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4d642fec-443e-4831-9047-95cc3a08d09f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.045198] env[61868]: DEBUG oslo_vmware.api [None req-40f8b9a6-0376-496a-97a7-807a4e018476 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315695, 'name': PowerOffVM_Task, 'duration_secs': 0.345838} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.045923] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-40f8b9a6-0376-496a-97a7-807a4e018476 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 929.046222] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-40f8b9a6-0376-496a-97a7-807a4e018476 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 929.046535] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fd9c6be1-f582-4844-8cc7-deafdac3cdc1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.051477] env[61868]: DEBUG oslo_vmware.api [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Waiting for the task: (returnval){ [ 929.051477] env[61868]: value = "task-1315696" [ 929.051477] env[61868]: _type = "Task" [ 929.051477] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.062670] env[61868]: DEBUG oslo_vmware.api [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': task-1315696, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.105059] env[61868]: DEBUG oslo_concurrency.lockutils [req-f1c13f2e-b1e5-43b0-b50d-b2fd18d435ac req-41e5706e-f327-4097-9090-b436db93a2e6 service nova] Releasing lock "refresh_cache-4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.105382] env[61868]: DEBUG nova.compute.manager [req-f1c13f2e-b1e5-43b0-b50d-b2fd18d435ac req-41e5706e-f327-4097-9090-b436db93a2e6 service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Received event network-changed-8a5f760c-c14e-4701-bde0-97fff3b6289e {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 929.105585] env[61868]: DEBUG nova.compute.manager [req-f1c13f2e-b1e5-43b0-b50d-b2fd18d435ac req-41e5706e-f327-4097-9090-b436db93a2e6 service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Refreshing instance network info cache due to event network-changed-8a5f760c-c14e-4701-bde0-97fff3b6289e. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 929.105820] env[61868]: DEBUG oslo_concurrency.lockutils [req-f1c13f2e-b1e5-43b0-b50d-b2fd18d435ac req-41e5706e-f327-4097-9090-b436db93a2e6 service nova] Acquiring lock "refresh_cache-5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.105960] env[61868]: DEBUG oslo_concurrency.lockutils [req-f1c13f2e-b1e5-43b0-b50d-b2fd18d435ac req-41e5706e-f327-4097-9090-b436db93a2e6 service nova] Acquired lock "refresh_cache-5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.106161] env[61868]: DEBUG nova.network.neutron [req-f1c13f2e-b1e5-43b0-b50d-b2fd18d435ac req-41e5706e-f327-4097-9090-b436db93a2e6 service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Refreshing network info cache for port 8a5f760c-c14e-4701-bde0-97fff3b6289e {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 929.120839] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-40f8b9a6-0376-496a-97a7-807a4e018476 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 929.121180] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-40f8b9a6-0376-496a-97a7-807a4e018476 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 929.121408] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-40f8b9a6-0376-496a-97a7-807a4e018476 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Deleting the datastore file [datastore1] 409cad1a-946d-4c58-aa57-1c0bf97fe63f {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 929.121668] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-749edd2c-7b6c-4430-b1d3-4f91d1ed2f4f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.127439] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-23a74488-062f-46a6-a299-a91a3d25288d tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 929.127817] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a151a997-29f1-4697-a962-584cdcf709e2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.132623] env[61868]: DEBUG oslo_vmware.api [None req-40f8b9a6-0376-496a-97a7-807a4e018476 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Waiting for the task: (returnval){ [ 929.132623] env[61868]: value = "task-1315698" [ 929.132623] env[61868]: _type = "Task" [ 929.132623] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.137380] env[61868]: DEBUG oslo_vmware.api [None req-23a74488-062f-46a6-a299-a91a3d25288d tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 929.137380] env[61868]: value = "task-1315699" [ 929.137380] env[61868]: _type = "Task" [ 929.137380] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.144642] env[61868]: DEBUG oslo_vmware.api [None req-40f8b9a6-0376-496a-97a7-807a4e018476 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315698, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.151602] env[61868]: DEBUG oslo_vmware.api [None req-23a74488-062f-46a6-a299-a91a3d25288d tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315699, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.202213] env[61868]: DEBUG oslo_vmware.api [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52ec8407-be98-f445-736c-5e7cc89bf113, 'name': SearchDatastore_Task, 'duration_secs': 0.011325} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.202804] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.203182] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 929.203516] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.203771] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.205250] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 929.205250] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4df37c17-c2a1-49f7-accb-38e35ba93f5b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.214455] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 929.214642] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 929.215313] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4803c6f1-720e-4e8b-9e3a-d7585a300701 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.221462] env[61868]: DEBUG oslo_vmware.api [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Waiting for the task: (returnval){ [ 929.221462] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]527e2357-f416-5319-c800-e536de4937b2" [ 929.221462] env[61868]: _type = "Task" [ 929.221462] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.231829] env[61868]: DEBUG oslo_vmware.api [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]527e2357-f416-5319-c800-e536de4937b2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.452819] env[61868]: DEBUG nova.scheduler.client.report [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 929.563123] env[61868]: DEBUG oslo_vmware.api [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': task-1315696, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.646572] env[61868]: DEBUG oslo_vmware.api [None req-40f8b9a6-0376-496a-97a7-807a4e018476 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Task: {'id': task-1315698, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.427598} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.647246] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-40f8b9a6-0376-496a-97a7-807a4e018476 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 929.647450] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-40f8b9a6-0376-496a-97a7-807a4e018476 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 929.647646] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-40f8b9a6-0376-496a-97a7-807a4e018476 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 929.647824] env[61868]: INFO nova.compute.manager [None req-40f8b9a6-0376-496a-97a7-807a4e018476 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Took 1.15 seconds to destroy the instance on the hypervisor. [ 929.648080] env[61868]: DEBUG oslo.service.loopingcall [None req-40f8b9a6-0376-496a-97a7-807a4e018476 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 929.648760] env[61868]: DEBUG nova.compute.manager [-] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 929.648760] env[61868]: DEBUG nova.network.neutron [-] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 929.653134] env[61868]: DEBUG oslo_vmware.api [None req-23a74488-062f-46a6-a299-a91a3d25288d tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315699, 'name': PowerOffVM_Task, 'duration_secs': 0.319199} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.653659] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-23a74488-062f-46a6-a299-a91a3d25288d tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 929.653847] env[61868]: DEBUG nova.compute.manager [None req-23a74488-062f-46a6-a299-a91a3d25288d tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 929.654815] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47112a4b-ed6b-4cb9-9ca1-53ba65b4d59c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.736171] env[61868]: DEBUG oslo_vmware.api [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]527e2357-f416-5319-c800-e536de4937b2, 'name': SearchDatastore_Task, 'duration_secs': 0.010211} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.737156] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a9ac4f26-e5e7-4b5b-9035-c468a3818231 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.746739] env[61868]: DEBUG oslo_vmware.api [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Waiting for the task: (returnval){ [ 929.746739] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52fd8ac1-13a4-6a49-529b-58f4da9986fc" [ 929.746739] env[61868]: _type = "Task" [ 929.746739] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.755625] env[61868]: DEBUG oslo_vmware.api [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52fd8ac1-13a4-6a49-529b-58f4da9986fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.959180] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.347s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.960114] env[61868]: DEBUG nova.compute.manager [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 929.966047] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 17.048s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.990221] env[61868]: DEBUG nova.network.neutron [req-f1c13f2e-b1e5-43b0-b50d-b2fd18d435ac req-41e5706e-f327-4097-9090-b436db93a2e6 service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Updated VIF entry in instance network info cache for port 8a5f760c-c14e-4701-bde0-97fff3b6289e. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 929.992173] env[61868]: DEBUG nova.network.neutron [req-f1c13f2e-b1e5-43b0-b50d-b2fd18d435ac req-41e5706e-f327-4097-9090-b436db93a2e6 service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Updating instance_info_cache with network_info: [{"id": "9771cedf-dae3-41e8-ad4b-aec421274d6c", "address": "fa:16:3e:4f:5f:b8", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.244", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9771cedf-da", "ovs_interfaceid": "9771cedf-dae3-41e8-ad4b-aec421274d6c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "8a5f760c-c14e-4701-bde0-97fff3b6289e", "address": "fa:16:3e:f7:7a:b3", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8a5f760c-c1", "ovs_interfaceid": "8a5f760c-c14e-4701-bde0-97fff3b6289e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.067808] env[61868]: DEBUG oslo_vmware.api [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': task-1315696, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.614388} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.068114] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] b51e968e-6911-47f3-b5a9-e8a5865e42a0/b51e968e-6911-47f3-b5a9-e8a5865e42a0.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 930.068345] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 930.068610] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-19e27e8d-11e0-4b51-8af4-17ab779cdd17 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.077515] env[61868]: DEBUG oslo_vmware.api [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Waiting for the task: (returnval){ [ 930.077515] env[61868]: value = "task-1315700" [ 930.077515] env[61868]: _type = "Task" [ 930.077515] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.086648] env[61868]: DEBUG oslo_vmware.api [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': task-1315700, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.167748] env[61868]: DEBUG oslo_concurrency.lockutils [None req-23a74488-062f-46a6-a299-a91a3d25288d tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "4e9d75e8-ca33-4e33-baa3-cb7d596f584e" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.559s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.259903] env[61868]: DEBUG oslo_vmware.api [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52fd8ac1-13a4-6a49-529b-58f4da9986fc, 'name': SearchDatastore_Task, 'duration_secs': 0.026091} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.260251] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.260524] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9/4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 930.261474] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ba492bc2-c63a-41e3-8234-28738814b057 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.272787] env[61868]: DEBUG oslo_vmware.api [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Waiting for the task: (returnval){ [ 930.272787] env[61868]: value = "task-1315701" [ 930.272787] env[61868]: _type = "Task" [ 930.272787] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.283122] env[61868]: DEBUG oslo_vmware.api [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Task: {'id': task-1315701, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.465510] env[61868]: DEBUG nova.compute.manager [req-b3db0ca2-85a0-4af9-b4a2-10fff903b015 req-f0707d21-6451-42c8-9904-e73357527e4b service nova] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Received event network-vif-deleted-29d18263-e917-46dc-8b91-33365c0d3481 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 930.467329] env[61868]: INFO nova.compute.manager [req-b3db0ca2-85a0-4af9-b4a2-10fff903b015 req-f0707d21-6451-42c8-9904-e73357527e4b service nova] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Neutron deleted interface 29d18263-e917-46dc-8b91-33365c0d3481; detaching it from the instance and deleting it from the info cache [ 930.467329] env[61868]: DEBUG nova.network.neutron [req-b3db0ca2-85a0-4af9-b4a2-10fff903b015 req-f0707d21-6451-42c8-9904-e73357527e4b service nova] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.474898] env[61868]: INFO nova.compute.claims [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 930.483473] env[61868]: DEBUG nova.compute.utils [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 930.490865] env[61868]: DEBUG nova.compute.manager [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 930.491154] env[61868]: DEBUG nova.network.neutron [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 930.494259] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c901db20-4fe2-4fd2-b265-a61f11905317 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.498722] env[61868]: DEBUG oslo_concurrency.lockutils [req-f1c13f2e-b1e5-43b0-b50d-b2fd18d435ac req-41e5706e-f327-4097-9090-b436db93a2e6 service nova] Releasing lock "refresh_cache-5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.512671] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f213b5c1-ed36-4d86-b8d5-a1d28b8c019f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.543335] env[61868]: DEBUG nova.policy [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '417d60bf01a24f01b7155a90c575aa4c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2947c69b0c484a0784451c7ec445cc3a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 930.558750] env[61868]: DEBUG nova.compute.manager [req-b3db0ca2-85a0-4af9-b4a2-10fff903b015 req-f0707d21-6451-42c8-9904-e73357527e4b service nova] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Detach interface failed, port_id=29d18263-e917-46dc-8b91-33365c0d3481, reason: Instance 409cad1a-946d-4c58-aa57-1c0bf97fe63f could not be found. {{(pid=61868) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 930.589171] env[61868]: DEBUG oslo_vmware.api [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': task-1315700, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.091582} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.589549] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 930.590879] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-069901db-ade1-4eec-8d2b-6ef51d7e42ef {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.615980] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] b51e968e-6911-47f3-b5a9-e8a5865e42a0/b51e968e-6911-47f3-b5a9-e8a5865e42a0.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 930.617182] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4d1e0b1d-462b-4b38-9fa1-8eda41ecc4e7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.642109] env[61868]: DEBUG oslo_vmware.api [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Waiting for the task: (returnval){ [ 930.642109] env[61868]: value = "task-1315702" [ 930.642109] env[61868]: _type = "Task" [ 930.642109] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.652506] env[61868]: DEBUG oslo_vmware.api [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': task-1315702, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.691435] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7f844117-d3b3-43c3-86a4-c26b82f8c10c tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "4e9d75e8-ca33-4e33-baa3-cb7d596f584e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.691757] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7f844117-d3b3-43c3-86a4-c26b82f8c10c tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "4e9d75e8-ca33-4e33-baa3-cb7d596f584e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.691998] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7f844117-d3b3-43c3-86a4-c26b82f8c10c tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "4e9d75e8-ca33-4e33-baa3-cb7d596f584e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.692247] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7f844117-d3b3-43c3-86a4-c26b82f8c10c tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "4e9d75e8-ca33-4e33-baa3-cb7d596f584e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.692456] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7f844117-d3b3-43c3-86a4-c26b82f8c10c tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "4e9d75e8-ca33-4e33-baa3-cb7d596f584e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.694875] env[61868]: INFO nova.compute.manager [None req-7f844117-d3b3-43c3-86a4-c26b82f8c10c tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Terminating instance [ 930.697084] env[61868]: DEBUG nova.compute.manager [None req-7f844117-d3b3-43c3-86a4-c26b82f8c10c tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 930.697324] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-7f844117-d3b3-43c3-86a4-c26b82f8c10c tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 930.698235] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09e3cb57-d972-48a5-8091-96c426dd6fc6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.707622] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-7f844117-d3b3-43c3-86a4-c26b82f8c10c tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 930.708088] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2c631f79-416d-4e28-aa50-119bad1d8a7f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.785258] env[61868]: DEBUG oslo_vmware.api [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Task: {'id': task-1315701, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.797026] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-7f844117-d3b3-43c3-86a4-c26b82f8c10c tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 930.797026] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-7f844117-d3b3-43c3-86a4-c26b82f8c10c tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 930.797026] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f844117-d3b3-43c3-86a4-c26b82f8c10c tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Deleting the datastore file [datastore1] 4e9d75e8-ca33-4e33-baa3-cb7d596f584e {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 930.797026] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ec68a11f-bc0d-4264-82e3-4313508a52d7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.807020] env[61868]: DEBUG oslo_vmware.api [None req-7f844117-d3b3-43c3-86a4-c26b82f8c10c tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 930.807020] env[61868]: value = "task-1315704" [ 930.807020] env[61868]: _type = "Task" [ 930.807020] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.815548] env[61868]: DEBUG oslo_vmware.api [None req-7f844117-d3b3-43c3-86a4-c26b82f8c10c tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315704, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.839167] env[61868]: DEBUG nova.network.neutron [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Successfully created port: 5ef47555-42a4-44dc-bf5f-01d21d921b5f {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 930.876617] env[61868]: DEBUG nova.network.neutron [-] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.987873] env[61868]: INFO nova.compute.resource_tracker [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Updating resource usage from migration 70dd0d99-e65a-47fd-872a-4c6b316b0d07 [ 930.999726] env[61868]: DEBUG nova.compute.manager [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 931.154759] env[61868]: DEBUG oslo_vmware.api [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': task-1315702, 'name': ReconfigVM_Task, 'duration_secs': 0.331414} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.155582] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Reconfigured VM instance instance-00000050 to attach disk [datastore1] b51e968e-6911-47f3-b5a9-e8a5865e42a0/b51e968e-6911-47f3-b5a9-e8a5865e42a0.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 931.155728] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9d647220-bd2d-43eb-ac1f-ebc1ebd054f3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.163966] env[61868]: DEBUG oslo_vmware.api [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Waiting for the task: (returnval){ [ 931.163966] env[61868]: value = "task-1315705" [ 931.163966] env[61868]: _type = "Task" [ 931.163966] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.173556] env[61868]: DEBUG oslo_vmware.api [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': task-1315705, 'name': Rename_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.238389] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b97a2721-6b9b-43c5-b09a-d03ac2a5f9c2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.246508] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-661eacd0-255f-486f-874e-e1f5551dbdf3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.279999] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59524aa7-832c-46b4-a5a7-0965f7c2edce {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.288087] env[61868]: DEBUG oslo_vmware.api [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Task: {'id': task-1315701, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.529821} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.290123] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9/4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 931.290361] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 931.290652] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-71e32e6c-ddd3-426d-9320-bb00fbadcc3b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.293420] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff656c85-7b5c-4337-8229-f1050643196a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.310462] env[61868]: DEBUG nova.compute.provider_tree [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 931.317860] env[61868]: DEBUG oslo_vmware.api [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Waiting for the task: (returnval){ [ 931.317860] env[61868]: value = "task-1315706" [ 931.317860] env[61868]: _type = "Task" [ 931.317860] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.326416] env[61868]: DEBUG oslo_vmware.api [None req-7f844117-d3b3-43c3-86a4-c26b82f8c10c tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315704, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.380827} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.327168] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f844117-d3b3-43c3-86a4-c26b82f8c10c tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 931.327382] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-7f844117-d3b3-43c3-86a4-c26b82f8c10c tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 931.327599] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-7f844117-d3b3-43c3-86a4-c26b82f8c10c tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 931.327745] env[61868]: INFO nova.compute.manager [None req-7f844117-d3b3-43c3-86a4-c26b82f8c10c tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Took 0.63 seconds to destroy the instance on the hypervisor. [ 931.328106] env[61868]: DEBUG oslo.service.loopingcall [None req-7f844117-d3b3-43c3-86a4-c26b82f8c10c tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 931.328586] env[61868]: DEBUG nova.compute.manager [-] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 931.328695] env[61868]: DEBUG nova.network.neutron [-] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 931.333938] env[61868]: DEBUG oslo_vmware.api [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Task: {'id': task-1315706, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.358809] env[61868]: DEBUG oslo_concurrency.lockutils [None req-dfc04811-68b9-46c1-b4dd-5650499adb22 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "interface-5aec2609-79d3-4725-a182-70b575adbe17-87455df2-50af-4a63-b58b-4813a20d0b38" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.359101] env[61868]: DEBUG oslo_concurrency.lockutils [None req-dfc04811-68b9-46c1-b4dd-5650499adb22 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "interface-5aec2609-79d3-4725-a182-70b575adbe17-87455df2-50af-4a63-b58b-4813a20d0b38" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.359477] env[61868]: DEBUG nova.objects.instance [None req-dfc04811-68b9-46c1-b4dd-5650499adb22 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lazy-loading 'flavor' on Instance uuid 5aec2609-79d3-4725-a182-70b575adbe17 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 931.379156] env[61868]: INFO nova.compute.manager [-] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Took 1.73 seconds to deallocate network for instance. [ 931.674521] env[61868]: DEBUG oslo_vmware.api [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': task-1315705, 'name': Rename_Task, 'duration_secs': 0.156698} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.674818] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 931.675090] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5931a790-0796-498b-8391-62cb07ecc426 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.683022] env[61868]: DEBUG oslo_vmware.api [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Waiting for the task: (returnval){ [ 931.683022] env[61868]: value = "task-1315707" [ 931.683022] env[61868]: _type = "Task" [ 931.683022] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.691234] env[61868]: DEBUG oslo_vmware.api [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': task-1315707, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.820223] env[61868]: DEBUG nova.scheduler.client.report [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 931.833266] env[61868]: DEBUG oslo_vmware.api [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Task: {'id': task-1315706, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061497} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.833849] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 931.834613] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bc095ea-f336-4cc0-80e4-d95c1d9792fa {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.858455] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9/4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 931.859123] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b75294a6-b409-45ba-99f5-d19f76724d41 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.882760] env[61868]: DEBUG oslo_vmware.api [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Waiting for the task: (returnval){ [ 931.882760] env[61868]: value = "task-1315708" [ 931.882760] env[61868]: _type = "Task" [ 931.882760] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.887221] env[61868]: DEBUG oslo_concurrency.lockutils [None req-40f8b9a6-0376-496a-97a7-807a4e018476 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.894362] env[61868]: DEBUG oslo_vmware.api [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Task: {'id': task-1315708, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.976939] env[61868]: DEBUG nova.objects.instance [None req-dfc04811-68b9-46c1-b4dd-5650499adb22 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lazy-loading 'pci_requests' on Instance uuid 5aec2609-79d3-4725-a182-70b575adbe17 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 932.012219] env[61868]: DEBUG nova.compute.manager [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 932.038208] env[61868]: DEBUG nova.virt.hardware [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 932.038476] env[61868]: DEBUG nova.virt.hardware [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 932.038640] env[61868]: DEBUG nova.virt.hardware [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 932.038827] env[61868]: DEBUG nova.virt.hardware [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 932.038978] env[61868]: DEBUG nova.virt.hardware [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 932.039154] env[61868]: DEBUG nova.virt.hardware [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 932.039368] env[61868]: DEBUG nova.virt.hardware [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 932.039535] env[61868]: DEBUG nova.virt.hardware [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 932.039708] env[61868]: DEBUG nova.virt.hardware [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 932.039879] env[61868]: DEBUG nova.virt.hardware [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 932.040072] env[61868]: DEBUG nova.virt.hardware [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 932.040951] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cde6cae-c30f-4acf-9662-b4722fcff0a6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.049536] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6db5a8b-7174-4b91-ae52-76fe51152c66 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.091014] env[61868]: DEBUG nova.network.neutron [-] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.195338] env[61868]: DEBUG oslo_vmware.api [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': task-1315707, 'name': PowerOnVM_Task, 'duration_secs': 0.450296} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.196131] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 932.196131] env[61868]: DEBUG nova.compute.manager [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 932.196653] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3cd0700-f7e8-4213-ab28-41d02ffc24fc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.325609] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.362s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.325834] env[61868]: INFO nova.compute.manager [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Migrating [ 932.326096] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.326241] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquired lock "compute-rpcapi-router" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.327567] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4752328e-63ce-4df8-ad78-8a6198a20895 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.178s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.327767] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4752328e-63ce-4df8-ad78-8a6198a20895 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.329894] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.443s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.331451] env[61868]: INFO nova.compute.claims [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 932.334892] env[61868]: INFO nova.compute.rpcapi [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 932.338765] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Releasing lock "compute-rpcapi-router" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.359142] env[61868]: INFO nova.scheduler.client.report [None req-4752328e-63ce-4df8-ad78-8a6198a20895 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Deleted allocations for instance f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3 [ 932.399984] env[61868]: DEBUG oslo_vmware.api [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Task: {'id': task-1315708, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.479535] env[61868]: DEBUG nova.objects.base [None req-dfc04811-68b9-46c1-b4dd-5650499adb22 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Object Instance<5aec2609-79d3-4725-a182-70b575adbe17> lazy-loaded attributes: flavor,pci_requests {{(pid=61868) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 932.479887] env[61868]: DEBUG nova.network.neutron [None req-dfc04811-68b9-46c1-b4dd-5650499adb22 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 932.557587] env[61868]: DEBUG nova.network.neutron [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Successfully updated port: 5ef47555-42a4-44dc-bf5f-01d21d921b5f {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 932.594076] env[61868]: INFO nova.compute.manager [-] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Took 1.27 seconds to deallocate network for instance. [ 932.618570] env[61868]: DEBUG nova.policy [None req-dfc04811-68b9-46c1-b4dd-5650499adb22 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c14b4ff01717495ca97c7f35f91c2995', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd7fc76299baf4a97b57139e5f1caa16e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 932.640390] env[61868]: DEBUG nova.compute.manager [req-1fb07e41-38bb-4aa0-a916-f0a8873218a6 req-dede1363-1aeb-4cb1-8873-a7acfb35d658 service nova] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Received event network-vif-deleted-ec859835-c58b-4c26-ac77-7628f73990b0 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 932.640536] env[61868]: DEBUG nova.compute.manager [req-1fb07e41-38bb-4aa0-a916-f0a8873218a6 req-dede1363-1aeb-4cb1-8873-a7acfb35d658 service nova] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Received event network-vif-plugged-5ef47555-42a4-44dc-bf5f-01d21d921b5f {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 932.641144] env[61868]: DEBUG oslo_concurrency.lockutils [req-1fb07e41-38bb-4aa0-a916-f0a8873218a6 req-dede1363-1aeb-4cb1-8873-a7acfb35d658 service nova] Acquiring lock "6ce39c07-939c-49f3-8871-6e2163709f4a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.641144] env[61868]: DEBUG oslo_concurrency.lockutils [req-1fb07e41-38bb-4aa0-a916-f0a8873218a6 req-dede1363-1aeb-4cb1-8873-a7acfb35d658 service nova] Lock "6ce39c07-939c-49f3-8871-6e2163709f4a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.641144] env[61868]: DEBUG oslo_concurrency.lockutils [req-1fb07e41-38bb-4aa0-a916-f0a8873218a6 req-dede1363-1aeb-4cb1-8873-a7acfb35d658 service nova] Lock "6ce39c07-939c-49f3-8871-6e2163709f4a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.641264] env[61868]: DEBUG nova.compute.manager [req-1fb07e41-38bb-4aa0-a916-f0a8873218a6 req-dede1363-1aeb-4cb1-8873-a7acfb35d658 service nova] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] No waiting events found dispatching network-vif-plugged-5ef47555-42a4-44dc-bf5f-01d21d921b5f {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 932.641626] env[61868]: WARNING nova.compute.manager [req-1fb07e41-38bb-4aa0-a916-f0a8873218a6 req-dede1363-1aeb-4cb1-8873-a7acfb35d658 service nova] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Received unexpected event network-vif-plugged-5ef47555-42a4-44dc-bf5f-01d21d921b5f for instance with vm_state building and task_state spawning. [ 932.714192] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.850879] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "refresh_cache-842c8606-632a-47d4-9a67-38d8e6ce6b4d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.850879] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquired lock "refresh_cache-842c8606-632a-47d4-9a67-38d8e6ce6b4d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.851177] env[61868]: DEBUG nova.network.neutron [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 932.868312] env[61868]: DEBUG oslo_concurrency.lockutils [None req-4752328e-63ce-4df8-ad78-8a6198a20895 tempest-ServerRescueTestJSONUnderV235-919206049 tempest-ServerRescueTestJSONUnderV235-919206049-project-member] Lock "f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.946s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.898844] env[61868]: DEBUG oslo_vmware.api [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Task: {'id': task-1315708, 'name': ReconfigVM_Task, 'duration_secs': 0.542125} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.899155] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Reconfigured VM instance instance-00000052 to attach disk [datastore1] 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9/4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 932.899846] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f9a4e33e-b327-4520-947b-9d1a7c9d7273 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.907832] env[61868]: DEBUG oslo_vmware.api [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Waiting for the task: (returnval){ [ 932.907832] env[61868]: value = "task-1315709" [ 932.907832] env[61868]: _type = "Task" [ 932.907832] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.916602] env[61868]: DEBUG oslo_vmware.api [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Task: {'id': task-1315709, 'name': Rename_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.061158] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Acquiring lock "refresh_cache-6ce39c07-939c-49f3-8871-6e2163709f4a" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.061214] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Acquired lock "refresh_cache-6ce39c07-939c-49f3-8871-6e2163709f4a" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.062233] env[61868]: DEBUG nova.network.neutron [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 933.100868] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7f844117-d3b3-43c3-86a4-c26b82f8c10c tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.303617] env[61868]: DEBUG oslo_concurrency.lockutils [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Acquiring lock "b51e968e-6911-47f3-b5a9-e8a5865e42a0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.303916] env[61868]: DEBUG oslo_concurrency.lockutils [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Lock "b51e968e-6911-47f3-b5a9-e8a5865e42a0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.304333] env[61868]: DEBUG oslo_concurrency.lockutils [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Acquiring lock "b51e968e-6911-47f3-b5a9-e8a5865e42a0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.304487] env[61868]: DEBUG oslo_concurrency.lockutils [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Lock "b51e968e-6911-47f3-b5a9-e8a5865e42a0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.304668] env[61868]: DEBUG oslo_concurrency.lockutils [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Lock "b51e968e-6911-47f3-b5a9-e8a5865e42a0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.307254] env[61868]: INFO nova.compute.manager [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Terminating instance [ 933.309358] env[61868]: DEBUG oslo_concurrency.lockutils [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Acquiring lock "refresh_cache-b51e968e-6911-47f3-b5a9-e8a5865e42a0" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.309580] env[61868]: DEBUG oslo_concurrency.lockutils [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Acquired lock "refresh_cache-b51e968e-6911-47f3-b5a9-e8a5865e42a0" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.309784] env[61868]: DEBUG nova.network.neutron [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 933.420759] env[61868]: DEBUG oslo_vmware.api [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Task: {'id': task-1315709, 'name': Rename_Task, 'duration_secs': 0.172913} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.421397] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 933.421397] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b47b736d-84fa-43bb-903b-4c9d97e4dd04 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.430356] env[61868]: DEBUG oslo_vmware.api [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Waiting for the task: (returnval){ [ 933.430356] env[61868]: value = "task-1315710" [ 933.430356] env[61868]: _type = "Task" [ 933.430356] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.439322] env[61868]: DEBUG oslo_vmware.api [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Task: {'id': task-1315710, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.596512] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c14e0750-9a15-4df2-b2fc-a631b3775e14 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.605566] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1187a429-cdb6-4376-b4f9-ac519103e7ea {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.638275] env[61868]: DEBUG nova.network.neutron [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 933.640471] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c49dbdec-c2b7-4249-b868-367716043a10 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.650478] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46773088-0799-4c8a-a6b3-dc0914b64a2f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.667626] env[61868]: DEBUG nova.compute.provider_tree [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 933.718864] env[61868]: DEBUG nova.network.neutron [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Updating instance_info_cache with network_info: [{"id": "e3943fa9-423d-4168-8aa6-1651280cd127", "address": "fa:16:3e:1d:68:a3", "network": {"id": "c1f5bf03-438f-4acc-93fd-159bed7d4069", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2061742146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a60c3c20950e4156b64c3b4c61b9f0f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24b91e3b-50f1-4a16-b929-942e6b31b2d8", "external-id": "nsx-vlan-transportzone-34", "segmentation_id": 34, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3943fa9-42", "ovs_interfaceid": "e3943fa9-423d-4168-8aa6-1651280cd127", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.838454] env[61868]: DEBUG nova.network.neutron [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Updating instance_info_cache with network_info: [{"id": "5ef47555-42a4-44dc-bf5f-01d21d921b5f", "address": "fa:16:3e:88:c5:ab", "network": {"id": "3cc88f4d-2861-44ab-8505-75bea36f5afa", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1992172748-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2947c69b0c484a0784451c7ec445cc3a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8459aaf-d6a8-46fb-ad14-464ac3104695", "external-id": "nsx-vlan-transportzone-46", "segmentation_id": 46, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ef47555-42", "ovs_interfaceid": "5ef47555-42a4-44dc-bf5f-01d21d921b5f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.838454] env[61868]: DEBUG nova.network.neutron [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 933.944333] env[61868]: DEBUG oslo_vmware.api [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Task: {'id': task-1315710, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.961723] env[61868]: DEBUG nova.network.neutron [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.171061] env[61868]: DEBUG nova.scheduler.client.report [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 934.222045] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Releasing lock "refresh_cache-842c8606-632a-47d4-9a67-38d8e6ce6b4d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.335772] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Releasing lock "refresh_cache-6ce39c07-939c-49f3-8871-6e2163709f4a" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.336155] env[61868]: DEBUG nova.compute.manager [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Instance network_info: |[{"id": "5ef47555-42a4-44dc-bf5f-01d21d921b5f", "address": "fa:16:3e:88:c5:ab", "network": {"id": "3cc88f4d-2861-44ab-8505-75bea36f5afa", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1992172748-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2947c69b0c484a0784451c7ec445cc3a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8459aaf-d6a8-46fb-ad14-464ac3104695", "external-id": "nsx-vlan-transportzone-46", "segmentation_id": 46, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ef47555-42", "ovs_interfaceid": "5ef47555-42a4-44dc-bf5f-01d21d921b5f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 934.336692] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:88:c5:ab', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c8459aaf-d6a8-46fb-ad14-464ac3104695', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5ef47555-42a4-44dc-bf5f-01d21d921b5f', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 934.348153] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Creating folder: Project (2947c69b0c484a0784451c7ec445cc3a). Parent ref: group-v281478. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 934.348524] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c91ca18b-544f-486f-b6fc-53c3804a3bec {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.362933] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Created folder: Project (2947c69b0c484a0784451c7ec445cc3a) in parent group-v281478. [ 934.362933] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Creating folder: Instances. Parent ref: group-v281582. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 934.364928] env[61868]: DEBUG nova.network.neutron [None req-dfc04811-68b9-46c1-b4dd-5650499adb22 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Successfully updated port: 87455df2-50af-4a63-b58b-4813a20d0b38 {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 934.366750] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ba42b5f7-cb0d-4e56-9cc8-2bad8557c3c0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.383719] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Created folder: Instances in parent group-v281582. [ 934.383719] env[61868]: DEBUG oslo.service.loopingcall [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 934.384139] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 934.384236] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-847411e5-e0d4-4d0a-acf7-840e492a7943 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.407463] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 934.407463] env[61868]: value = "task-1315713" [ 934.407463] env[61868]: _type = "Task" [ 934.407463] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.418714] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315713, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.442797] env[61868]: DEBUG oslo_vmware.api [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Task: {'id': task-1315710, 'name': PowerOnVM_Task, 'duration_secs': 0.587749} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.443160] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 934.443390] env[61868]: INFO nova.compute.manager [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Took 10.16 seconds to spawn the instance on the hypervisor. [ 934.443657] env[61868]: DEBUG nova.compute.manager [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 934.444493] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b01bcc3a-8395-4ab8-b231-9024bc64d7f3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.467145] env[61868]: DEBUG oslo_concurrency.lockutils [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Releasing lock "refresh_cache-b51e968e-6911-47f3-b5a9-e8a5865e42a0" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.467668] env[61868]: DEBUG nova.compute.manager [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 934.467925] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 934.469021] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43bd4dd8-c9a4-4187-b871-c58d5da01473 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.480521] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 934.481532] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-adea45c1-ff40-4173-a1fc-8a9d9a8e0e95 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.491058] env[61868]: DEBUG oslo_vmware.api [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Waiting for the task: (returnval){ [ 934.491058] env[61868]: value = "task-1315714" [ 934.491058] env[61868]: _type = "Task" [ 934.491058] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.502827] env[61868]: DEBUG oslo_vmware.api [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': task-1315714, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.682106] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.352s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.682839] env[61868]: DEBUG nova.compute.manager [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 934.685738] env[61868]: DEBUG oslo_concurrency.lockutils [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.080s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.687862] env[61868]: INFO nova.compute.claims [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 934.872565] env[61868]: DEBUG oslo_concurrency.lockutils [None req-dfc04811-68b9-46c1-b4dd-5650499adb22 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "refresh_cache-5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.872565] env[61868]: DEBUG oslo_concurrency.lockutils [None req-dfc04811-68b9-46c1-b4dd-5650499adb22 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquired lock "refresh_cache-5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.872565] env[61868]: DEBUG nova.network.neutron [None req-dfc04811-68b9-46c1-b4dd-5650499adb22 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 934.918030] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315713, 'name': CreateVM_Task, 'duration_secs': 0.33813} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.918574] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 934.918870] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.919060] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.919661] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 934.920028] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-594d2b78-f679-4858-9595-fbff3406d09b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.925930] env[61868]: DEBUG oslo_vmware.api [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Waiting for the task: (returnval){ [ 934.925930] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52173dd2-6448-db9a-ff9d-f6671e8955ff" [ 934.925930] env[61868]: _type = "Task" [ 934.925930] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.934566] env[61868]: DEBUG oslo_vmware.api [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52173dd2-6448-db9a-ff9d-f6671e8955ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.962153] env[61868]: INFO nova.compute.manager [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Took 44.27 seconds to build instance. [ 934.966123] env[61868]: DEBUG nova.compute.manager [req-f85ceb5e-13cc-4d2b-8b8e-f2a4b405b757 req-0d9150e0-5c4e-4ff5-a623-be5418384d31 service nova] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Received event network-changed-5ef47555-42a4-44dc-bf5f-01d21d921b5f {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 934.966296] env[61868]: DEBUG nova.compute.manager [req-f85ceb5e-13cc-4d2b-8b8e-f2a4b405b757 req-0d9150e0-5c4e-4ff5-a623-be5418384d31 service nova] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Refreshing instance network info cache due to event network-changed-5ef47555-42a4-44dc-bf5f-01d21d921b5f. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 934.966486] env[61868]: DEBUG oslo_concurrency.lockutils [req-f85ceb5e-13cc-4d2b-8b8e-f2a4b405b757 req-0d9150e0-5c4e-4ff5-a623-be5418384d31 service nova] Acquiring lock "refresh_cache-6ce39c07-939c-49f3-8871-6e2163709f4a" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.966737] env[61868]: DEBUG oslo_concurrency.lockutils [req-f85ceb5e-13cc-4d2b-8b8e-f2a4b405b757 req-0d9150e0-5c4e-4ff5-a623-be5418384d31 service nova] Acquired lock "refresh_cache-6ce39c07-939c-49f3-8871-6e2163709f4a" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.966961] env[61868]: DEBUG nova.network.neutron [req-f85ceb5e-13cc-4d2b-8b8e-f2a4b405b757 req-0d9150e0-5c4e-4ff5-a623-be5418384d31 service nova] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Refreshing network info cache for port 5ef47555-42a4-44dc-bf5f-01d21d921b5f {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 935.003116] env[61868]: DEBUG oslo_vmware.api [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': task-1315714, 'name': PowerOffVM_Task, 'duration_secs': 0.191952} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.003673] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 935.003960] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 935.004379] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-65560f0a-d497-402d-b233-c2e0d29a4882 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.034329] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 935.034329] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 935.034329] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Deleting the datastore file [datastore1] b51e968e-6911-47f3-b5a9-e8a5865e42a0 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 935.034329] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6b241746-081b-4f51-9109-2fffe449aee8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.044409] env[61868]: DEBUG oslo_vmware.api [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Waiting for the task: (returnval){ [ 935.044409] env[61868]: value = "task-1315716" [ 935.044409] env[61868]: _type = "Task" [ 935.044409] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.055255] env[61868]: DEBUG oslo_vmware.api [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': task-1315716, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.193038] env[61868]: DEBUG nova.compute.utils [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 935.196045] env[61868]: DEBUG nova.compute.manager [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 935.196282] env[61868]: DEBUG nova.network.neutron [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 935.235519] env[61868]: DEBUG nova.policy [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dc7f7945969547b8ba739e7f9ef0a9b8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0880f172b6534c92be8ba18e6cab17fe', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 935.431066] env[61868]: WARNING nova.network.neutron [None req-dfc04811-68b9-46c1-b4dd-5650499adb22 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] d4f100cd-8de6-4459-a7a6-bc1895b8b14f already exists in list: networks containing: ['d4f100cd-8de6-4459-a7a6-bc1895b8b14f']. ignoring it [ 935.431348] env[61868]: WARNING nova.network.neutron [None req-dfc04811-68b9-46c1-b4dd-5650499adb22 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] d4f100cd-8de6-4459-a7a6-bc1895b8b14f already exists in list: networks containing: ['d4f100cd-8de6-4459-a7a6-bc1895b8b14f']. ignoring it [ 935.439634] env[61868]: DEBUG oslo_vmware.api [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52173dd2-6448-db9a-ff9d-f6671e8955ff, 'name': SearchDatastore_Task, 'duration_secs': 0.009768} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.440033] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.440344] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 935.440640] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.440837] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.441128] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 935.441479] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-91f356af-9e88-4bb5-b898-c6cbbe0ef364 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.453291] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 935.453692] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 935.454494] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3bac3c7-0cb5-4c47-a678-78e00876bc92 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.460557] env[61868]: DEBUG oslo_vmware.api [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Waiting for the task: (returnval){ [ 935.460557] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5268dff8-6345-d1d8-037e-f75f6233f048" [ 935.460557] env[61868]: _type = "Task" [ 935.460557] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.464099] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ceb30edd-3049-4acf-864b-59a4b3d9a57e tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Lock "4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.787s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.471899] env[61868]: DEBUG oslo_vmware.api [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5268dff8-6345-d1d8-037e-f75f6233f048, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.559979] env[61868]: DEBUG oslo_vmware.api [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Task: {'id': task-1315716, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.104145} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.560472] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 935.560750] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 935.560986] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 935.561380] env[61868]: INFO nova.compute.manager [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Took 1.09 seconds to destroy the instance on the hypervisor. [ 935.561921] env[61868]: DEBUG oslo.service.loopingcall [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 935.562463] env[61868]: DEBUG nova.compute.manager [-] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 935.562630] env[61868]: DEBUG nova.network.neutron [-] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 935.598695] env[61868]: DEBUG nova.network.neutron [-] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 935.636186] env[61868]: DEBUG nova.network.neutron [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Successfully created port: bbcb967e-bf16-4589-ab6a-23e5ce9e8dd4 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 935.704334] env[61868]: DEBUG nova.compute.manager [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 935.745821] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc5d7df5-a426-43ad-8dc4-217a94750df1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.768250] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Updating instance '842c8606-632a-47d4-9a67-38d8e6ce6b4d' progress to 0 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 935.773036] env[61868]: DEBUG nova.network.neutron [req-f85ceb5e-13cc-4d2b-8b8e-f2a4b405b757 req-0d9150e0-5c4e-4ff5-a623-be5418384d31 service nova] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Updated VIF entry in instance network info cache for port 5ef47555-42a4-44dc-bf5f-01d21d921b5f. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 935.773550] env[61868]: DEBUG nova.network.neutron [req-f85ceb5e-13cc-4d2b-8b8e-f2a4b405b757 req-0d9150e0-5c4e-4ff5-a623-be5418384d31 service nova] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Updating instance_info_cache with network_info: [{"id": "5ef47555-42a4-44dc-bf5f-01d21d921b5f", "address": "fa:16:3e:88:c5:ab", "network": {"id": "3cc88f4d-2861-44ab-8505-75bea36f5afa", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1992172748-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2947c69b0c484a0784451c7ec445cc3a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8459aaf-d6a8-46fb-ad14-464ac3104695", "external-id": "nsx-vlan-transportzone-46", "segmentation_id": 46, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ef47555-42", "ovs_interfaceid": "5ef47555-42a4-44dc-bf5f-01d21d921b5f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.974761] env[61868]: DEBUG oslo_vmware.api [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5268dff8-6345-d1d8-037e-f75f6233f048, 'name': SearchDatastore_Task, 'duration_secs': 0.021633} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.975686] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-651909d4-6355-48ae-a737-cb35da4f475e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.981473] env[61868]: DEBUG oslo_vmware.api [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Waiting for the task: (returnval){ [ 935.981473] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5214c8cd-8882-ef8a-c51f-c12b1b94b02c" [ 935.981473] env[61868]: _type = "Task" [ 935.981473] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.988619] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39de9ed3-bdc6-4ff0-83dc-3e2149aa901a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.000350] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-071fce98-5fac-4043-9431-fa0f3a0502a8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.004087] env[61868]: DEBUG oslo_vmware.api [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5214c8cd-8882-ef8a-c51f-c12b1b94b02c, 'name': SearchDatastore_Task, 'duration_secs': 0.010247} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.004401] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.004712] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 6ce39c07-939c-49f3-8871-6e2163709f4a/6ce39c07-939c-49f3-8871-6e2163709f4a.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 936.005363] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6aa1a26d-c438-470e-9b29-cd352ad5bf4b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.036706] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edf29e43-2260-4152-be90-fedd525666f3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.039593] env[61868]: DEBUG oslo_vmware.api [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Waiting for the task: (returnval){ [ 936.039593] env[61868]: value = "task-1315717" [ 936.039593] env[61868]: _type = "Task" [ 936.039593] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.049647] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94fb0edb-9ca0-4e60-9967-19cd0d7e3e83 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.056895] env[61868]: DEBUG oslo_vmware.api [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Task: {'id': task-1315717, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.067618] env[61868]: DEBUG nova.compute.provider_tree [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 936.101438] env[61868]: DEBUG nova.network.neutron [-] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.164297] env[61868]: DEBUG nova.network.neutron [None req-dfc04811-68b9-46c1-b4dd-5650499adb22 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Updating instance_info_cache with network_info: [{"id": "9771cedf-dae3-41e8-ad4b-aec421274d6c", "address": "fa:16:3e:4f:5f:b8", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.244", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9771cedf-da", "ovs_interfaceid": "9771cedf-dae3-41e8-ad4b-aec421274d6c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "8a5f760c-c14e-4701-bde0-97fff3b6289e", "address": "fa:16:3e:f7:7a:b3", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8a5f760c-c1", "ovs_interfaceid": "8a5f760c-c14e-4701-bde0-97fff3b6289e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "87455df2-50af-4a63-b58b-4813a20d0b38", "address": "fa:16:3e:51:bc:03", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87455df2-50", "ovs_interfaceid": "87455df2-50af-4a63-b58b-4813a20d0b38", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.279756] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 936.280355] env[61868]: DEBUG oslo_concurrency.lockutils [req-f85ceb5e-13cc-4d2b-8b8e-f2a4b405b757 req-0d9150e0-5c4e-4ff5-a623-be5418384d31 service nova] Releasing lock "refresh_cache-6ce39c07-939c-49f3-8871-6e2163709f4a" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.280611] env[61868]: DEBUG nova.compute.manager [req-f85ceb5e-13cc-4d2b-8b8e-f2a4b405b757 req-0d9150e0-5c4e-4ff5-a623-be5418384d31 service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Received event network-vif-plugged-87455df2-50af-4a63-b58b-4813a20d0b38 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 936.280805] env[61868]: DEBUG oslo_concurrency.lockutils [req-f85ceb5e-13cc-4d2b-8b8e-f2a4b405b757 req-0d9150e0-5c4e-4ff5-a623-be5418384d31 service nova] Acquiring lock "5aec2609-79d3-4725-a182-70b575adbe17-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.281025] env[61868]: DEBUG oslo_concurrency.lockutils [req-f85ceb5e-13cc-4d2b-8b8e-f2a4b405b757 req-0d9150e0-5c4e-4ff5-a623-be5418384d31 service nova] Lock "5aec2609-79d3-4725-a182-70b575adbe17-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.281197] env[61868]: DEBUG oslo_concurrency.lockutils [req-f85ceb5e-13cc-4d2b-8b8e-f2a4b405b757 req-0d9150e0-5c4e-4ff5-a623-be5418384d31 service nova] Lock "5aec2609-79d3-4725-a182-70b575adbe17-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.281366] env[61868]: DEBUG nova.compute.manager [req-f85ceb5e-13cc-4d2b-8b8e-f2a4b405b757 req-0d9150e0-5c4e-4ff5-a623-be5418384d31 service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] No waiting events found dispatching network-vif-plugged-87455df2-50af-4a63-b58b-4813a20d0b38 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 936.281536] env[61868]: WARNING nova.compute.manager [req-f85ceb5e-13cc-4d2b-8b8e-f2a4b405b757 req-0d9150e0-5c4e-4ff5-a623-be5418384d31 service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Received unexpected event network-vif-plugged-87455df2-50af-4a63-b58b-4813a20d0b38 for instance with vm_state active and task_state None. [ 936.281699] env[61868]: DEBUG nova.compute.manager [req-f85ceb5e-13cc-4d2b-8b8e-f2a4b405b757 req-0d9150e0-5c4e-4ff5-a623-be5418384d31 service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Received event network-changed-87455df2-50af-4a63-b58b-4813a20d0b38 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 936.281856] env[61868]: DEBUG nova.compute.manager [req-f85ceb5e-13cc-4d2b-8b8e-f2a4b405b757 req-0d9150e0-5c4e-4ff5-a623-be5418384d31 service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Refreshing instance network info cache due to event network-changed-87455df2-50af-4a63-b58b-4813a20d0b38. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 936.282333] env[61868]: DEBUG oslo_concurrency.lockutils [req-f85ceb5e-13cc-4d2b-8b8e-f2a4b405b757 req-0d9150e0-5c4e-4ff5-a623-be5418384d31 service nova] Acquiring lock "refresh_cache-5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.282333] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-63d24698-8744-4906-a79a-851767cef071 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.291451] env[61868]: DEBUG oslo_vmware.api [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 936.291451] env[61868]: value = "task-1315718" [ 936.291451] env[61868]: _type = "Task" [ 936.291451] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.303285] env[61868]: DEBUG oslo_vmware.api [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315718, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.550310] env[61868]: DEBUG oslo_vmware.api [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Task: {'id': task-1315717, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.515981} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.550580] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 6ce39c07-939c-49f3-8871-6e2163709f4a/6ce39c07-939c-49f3-8871-6e2163709f4a.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 936.550791] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 936.551055] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bb565cf1-b93d-417e-9f74-fd9e7a065171 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.557622] env[61868]: DEBUG oslo_vmware.api [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Waiting for the task: (returnval){ [ 936.557622] env[61868]: value = "task-1315719" [ 936.557622] env[61868]: _type = "Task" [ 936.557622] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.566545] env[61868]: DEBUG oslo_vmware.api [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Task: {'id': task-1315719, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.571551] env[61868]: DEBUG nova.scheduler.client.report [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 936.603936] env[61868]: INFO nova.compute.manager [-] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Took 1.04 seconds to deallocate network for instance. [ 936.668088] env[61868]: DEBUG oslo_concurrency.lockutils [None req-dfc04811-68b9-46c1-b4dd-5650499adb22 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Releasing lock "refresh_cache-5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.669167] env[61868]: DEBUG oslo_concurrency.lockutils [None req-dfc04811-68b9-46c1-b4dd-5650499adb22 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.669167] env[61868]: DEBUG oslo_concurrency.lockutils [None req-dfc04811-68b9-46c1-b4dd-5650499adb22 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquired lock "5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.669342] env[61868]: DEBUG oslo_concurrency.lockutils [req-f85ceb5e-13cc-4d2b-8b8e-f2a4b405b757 req-0d9150e0-5c4e-4ff5-a623-be5418384d31 service nova] Acquired lock "refresh_cache-5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.669487] env[61868]: DEBUG nova.network.neutron [req-f85ceb5e-13cc-4d2b-8b8e-f2a4b405b757 req-0d9150e0-5c4e-4ff5-a623-be5418384d31 service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Refreshing network info cache for port 87455df2-50af-4a63-b58b-4813a20d0b38 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 936.671206] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6a14c81-9362-44b7-b57a-a9d04d278e41 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.690241] env[61868]: DEBUG nova.virt.hardware [None req-dfc04811-68b9-46c1-b4dd-5650499adb22 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 936.690506] env[61868]: DEBUG nova.virt.hardware [None req-dfc04811-68b9-46c1-b4dd-5650499adb22 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 936.690669] env[61868]: DEBUG nova.virt.hardware [None req-dfc04811-68b9-46c1-b4dd-5650499adb22 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 936.690853] env[61868]: DEBUG nova.virt.hardware [None req-dfc04811-68b9-46c1-b4dd-5650499adb22 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 936.691030] env[61868]: DEBUG nova.virt.hardware [None req-dfc04811-68b9-46c1-b4dd-5650499adb22 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 936.691224] env[61868]: DEBUG nova.virt.hardware [None req-dfc04811-68b9-46c1-b4dd-5650499adb22 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 936.691478] env[61868]: DEBUG nova.virt.hardware [None req-dfc04811-68b9-46c1-b4dd-5650499adb22 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 936.691678] env[61868]: DEBUG nova.virt.hardware [None req-dfc04811-68b9-46c1-b4dd-5650499adb22 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 936.691857] env[61868]: DEBUG nova.virt.hardware [None req-dfc04811-68b9-46c1-b4dd-5650499adb22 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 936.692057] env[61868]: DEBUG nova.virt.hardware [None req-dfc04811-68b9-46c1-b4dd-5650499adb22 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 936.692253] env[61868]: DEBUG nova.virt.hardware [None req-dfc04811-68b9-46c1-b4dd-5650499adb22 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 936.698819] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-dfc04811-68b9-46c1-b4dd-5650499adb22 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Reconfiguring VM to attach interface {{(pid=61868) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 936.699703] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4e8cd27f-0246-42fc-90d5-873a689dbca8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.718583] env[61868]: DEBUG oslo_vmware.api [None req-dfc04811-68b9-46c1-b4dd-5650499adb22 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for the task: (returnval){ [ 936.718583] env[61868]: value = "task-1315720" [ 936.718583] env[61868]: _type = "Task" [ 936.718583] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.725072] env[61868]: DEBUG nova.compute.manager [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 936.730500] env[61868]: DEBUG oslo_vmware.api [None req-dfc04811-68b9-46c1-b4dd-5650499adb22 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315720, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.755397] env[61868]: DEBUG nova.virt.hardware [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 936.755684] env[61868]: DEBUG nova.virt.hardware [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 936.755873] env[61868]: DEBUG nova.virt.hardware [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 936.756174] env[61868]: DEBUG nova.virt.hardware [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 936.756296] env[61868]: DEBUG nova.virt.hardware [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 936.756568] env[61868]: DEBUG nova.virt.hardware [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 936.756792] env[61868]: DEBUG nova.virt.hardware [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 936.756978] env[61868]: DEBUG nova.virt.hardware [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 936.757223] env[61868]: DEBUG nova.virt.hardware [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 936.757351] env[61868]: DEBUG nova.virt.hardware [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 936.757533] env[61868]: DEBUG nova.virt.hardware [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 936.758415] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a6a2a0b-6eef-4078-be09-ce18b2695902 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.768180] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec6aab98-21c6-4f02-8d2f-0c3e36be06c3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.803779] env[61868]: DEBUG oslo_vmware.api [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315718, 'name': PowerOffVM_Task, 'duration_secs': 0.21759} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.804186] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 936.804486] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Updating instance '842c8606-632a-47d4-9a67-38d8e6ce6b4d' progress to 17 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 937.067895] env[61868]: DEBUG oslo_vmware.api [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Task: {'id': task-1315719, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074937} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.069025] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 937.069552] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d43dfa89-4409-427a-b39e-3fbb88f56492 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.084071] env[61868]: DEBUG oslo_concurrency.lockutils [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.398s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.084603] env[61868]: DEBUG nova.compute.manager [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 937.095892] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] 6ce39c07-939c-49f3-8871-6e2163709f4a/6ce39c07-939c-49f3-8871-6e2163709f4a.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 937.096544] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1f690bc9-2e92-4774-a822-25d8505bdac3 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.018s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.096747] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1f690bc9-2e92-4774-a822-25d8505bdac3 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.099134] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e5888991-048d-4aa8-b4bd-2d8f6de4700f tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.771s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.099341] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e5888991-048d-4aa8-b4bd-2d8f6de4700f tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.101201] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a29eb12d-21de-449b-99db-b45133966326 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.685s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.101438] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a29eb12d-21de-449b-99db-b45133966326 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.104712] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.980s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.105826] env[61868]: INFO nova.compute.claims [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 937.108396] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a929d0b5-d343-4660-b435-0d69187347d2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.124801] env[61868]: DEBUG oslo_concurrency.lockutils [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.137080] env[61868]: DEBUG oslo_vmware.api [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Waiting for the task: (returnval){ [ 937.137080] env[61868]: value = "task-1315721" [ 937.137080] env[61868]: _type = "Task" [ 937.137080] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.152206] env[61868]: INFO nova.scheduler.client.report [None req-e5888991-048d-4aa8-b4bd-2d8f6de4700f tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Deleted allocations for instance 306f2dc4-933c-4948-8330-eda93931c4e0 [ 937.154564] env[61868]: INFO nova.scheduler.client.report [None req-a29eb12d-21de-449b-99db-b45133966326 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Deleted allocations for instance 8ea8e28b-3582-45f2-b4b5-84f624415a58 [ 937.159599] env[61868]: DEBUG oslo_vmware.api [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Task: {'id': task-1315721, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.234809] env[61868]: DEBUG oslo_vmware.api [None req-dfc04811-68b9-46c1-b4dd-5650499adb22 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315720, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.311982] env[61868]: DEBUG nova.virt.hardware [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 937.312408] env[61868]: DEBUG nova.virt.hardware [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 937.312603] env[61868]: DEBUG nova.virt.hardware [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 937.312797] env[61868]: DEBUG nova.virt.hardware [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 937.312957] env[61868]: DEBUG nova.virt.hardware [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 937.313421] env[61868]: DEBUG nova.virt.hardware [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 937.313739] env[61868]: DEBUG nova.virt.hardware [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 937.314266] env[61868]: DEBUG nova.virt.hardware [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 937.314606] env[61868]: DEBUG nova.virt.hardware [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 937.314801] env[61868]: DEBUG nova.virt.hardware [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 937.315031] env[61868]: DEBUG nova.virt.hardware [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 937.320165] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e2d38b0c-4d02-4b74-ba94-f7104aebd7d7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.342462] env[61868]: DEBUG oslo_vmware.api [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 937.342462] env[61868]: value = "task-1315722" [ 937.342462] env[61868]: _type = "Task" [ 937.342462] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.354009] env[61868]: DEBUG oslo_vmware.api [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315722, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.531976] env[61868]: DEBUG nova.network.neutron [req-f85ceb5e-13cc-4d2b-8b8e-f2a4b405b757 req-0d9150e0-5c4e-4ff5-a623-be5418384d31 service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Updated VIF entry in instance network info cache for port 87455df2-50af-4a63-b58b-4813a20d0b38. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 937.532885] env[61868]: DEBUG nova.network.neutron [req-f85ceb5e-13cc-4d2b-8b8e-f2a4b405b757 req-0d9150e0-5c4e-4ff5-a623-be5418384d31 service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Updating instance_info_cache with network_info: [{"id": "9771cedf-dae3-41e8-ad4b-aec421274d6c", "address": "fa:16:3e:4f:5f:b8", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.244", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9771cedf-da", "ovs_interfaceid": "9771cedf-dae3-41e8-ad4b-aec421274d6c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "8a5f760c-c14e-4701-bde0-97fff3b6289e", "address": "fa:16:3e:f7:7a:b3", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8a5f760c-c1", "ovs_interfaceid": "8a5f760c-c14e-4701-bde0-97fff3b6289e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "87455df2-50af-4a63-b58b-4813a20d0b38", "address": "fa:16:3e:51:bc:03", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87455df2-50", "ovs_interfaceid": "87455df2-50af-4a63-b58b-4813a20d0b38", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.572055] env[61868]: DEBUG nova.network.neutron [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Successfully updated port: bbcb967e-bf16-4589-ab6a-23e5ce9e8dd4 {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 937.575958] env[61868]: DEBUG nova.compute.manager [req-83e350c0-bee2-4722-b78a-2fed48eb0ade req-b331199a-4594-4dfc-9de8-5d286c7be0eb service nova] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Received event network-vif-plugged-bbcb967e-bf16-4589-ab6a-23e5ce9e8dd4 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 937.576201] env[61868]: DEBUG oslo_concurrency.lockutils [req-83e350c0-bee2-4722-b78a-2fed48eb0ade req-b331199a-4594-4dfc-9de8-5d286c7be0eb service nova] Acquiring lock "87b25494-1853-4c7b-ae8b-6283992e895c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.576731] env[61868]: DEBUG oslo_concurrency.lockutils [req-83e350c0-bee2-4722-b78a-2fed48eb0ade req-b331199a-4594-4dfc-9de8-5d286c7be0eb service nova] Lock "87b25494-1853-4c7b-ae8b-6283992e895c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.576967] env[61868]: DEBUG oslo_concurrency.lockutils [req-83e350c0-bee2-4722-b78a-2fed48eb0ade req-b331199a-4594-4dfc-9de8-5d286c7be0eb service nova] Lock "87b25494-1853-4c7b-ae8b-6283992e895c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.577179] env[61868]: DEBUG nova.compute.manager [req-83e350c0-bee2-4722-b78a-2fed48eb0ade req-b331199a-4594-4dfc-9de8-5d286c7be0eb service nova] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] No waiting events found dispatching network-vif-plugged-bbcb967e-bf16-4589-ab6a-23e5ce9e8dd4 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 937.577325] env[61868]: WARNING nova.compute.manager [req-83e350c0-bee2-4722-b78a-2fed48eb0ade req-b331199a-4594-4dfc-9de8-5d286c7be0eb service nova] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Received unexpected event network-vif-plugged-bbcb967e-bf16-4589-ab6a-23e5ce9e8dd4 for instance with vm_state building and task_state spawning. [ 937.597822] env[61868]: DEBUG nova.compute.utils [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 937.600051] env[61868]: DEBUG nova.compute.manager [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 937.600051] env[61868]: DEBUG nova.network.neutron [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 937.641646] env[61868]: DEBUG nova.policy [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0b615dd3d1064706bea31b52fbfa9a61', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5658848e1b0e42929c04a4a8de40a291', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 937.651639] env[61868]: DEBUG oslo_vmware.api [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Task: {'id': task-1315721, 'name': ReconfigVM_Task, 'duration_secs': 0.327051} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.652142] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1f690bc9-2e92-4774-a822-25d8505bdac3 tempest-AttachVolumeShelveTestJSON-1432085417 tempest-AttachVolumeShelveTestJSON-1432085417-project-member] Lock "083140bf-233d-49ac-8247-cd93206f9200" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.078s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.653317] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Reconfigured VM instance instance-00000053 to attach disk [datastore1] 6ce39c07-939c-49f3-8871-6e2163709f4a/6ce39c07-939c-49f3-8871-6e2163709f4a.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 937.654106] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c805e8b6-d01b-44b5-b397-79010f5ffafa {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.661365] env[61868]: DEBUG oslo_vmware.api [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Waiting for the task: (returnval){ [ 937.661365] env[61868]: value = "task-1315723" [ 937.661365] env[61868]: _type = "Task" [ 937.661365] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.672436] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a29eb12d-21de-449b-99db-b45133966326 tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Lock "8ea8e28b-3582-45f2-b4b5-84f624415a58" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.388s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.675580] env[61868]: DEBUG oslo_vmware.api [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Task: {'id': task-1315723, 'name': Rename_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.676202] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e5888991-048d-4aa8-b4bd-2d8f6de4700f tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "306f2dc4-933c-4948-8330-eda93931c4e0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.310s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.731185] env[61868]: DEBUG oslo_vmware.api [None req-dfc04811-68b9-46c1-b4dd-5650499adb22 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315720, 'name': ReconfigVM_Task, 'duration_secs': 0.578187} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.731771] env[61868]: DEBUG oslo_concurrency.lockutils [None req-dfc04811-68b9-46c1-b4dd-5650499adb22 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Releasing lock "5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.732055] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-dfc04811-68b9-46c1-b4dd-5650499adb22 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Reconfigured VM to attach interface {{(pid=61868) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 937.855898] env[61868]: DEBUG oslo_vmware.api [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315722, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.897662] env[61868]: DEBUG nova.network.neutron [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Successfully created port: 1f37fd44-6543-44b3-9fce-cb4154f0f0d0 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 938.036452] env[61868]: DEBUG oslo_concurrency.lockutils [req-f85ceb5e-13cc-4d2b-8b8e-f2a4b405b757 req-0d9150e0-5c4e-4ff5-a623-be5418384d31 service nova] Releasing lock "refresh_cache-5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.052298] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d54e0102-d52b-44c3-93dd-ba46528a03fc tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Acquiring lock "19c2720c-90bc-47f6-999b-6031f893408d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.052574] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d54e0102-d52b-44c3-93dd-ba46528a03fc tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Lock "19c2720c-90bc-47f6-999b-6031f893408d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.052791] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d54e0102-d52b-44c3-93dd-ba46528a03fc tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Acquiring lock "19c2720c-90bc-47f6-999b-6031f893408d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.053034] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d54e0102-d52b-44c3-93dd-ba46528a03fc tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Lock "19c2720c-90bc-47f6-999b-6031f893408d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.053278] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d54e0102-d52b-44c3-93dd-ba46528a03fc tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Lock "19c2720c-90bc-47f6-999b-6031f893408d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.055448] env[61868]: INFO nova.compute.manager [None req-d54e0102-d52b-44c3-93dd-ba46528a03fc tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Terminating instance [ 938.057962] env[61868]: DEBUG nova.compute.manager [None req-d54e0102-d52b-44c3-93dd-ba46528a03fc tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 938.058203] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-d54e0102-d52b-44c3-93dd-ba46528a03fc tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 938.059044] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5abf8fb-d7a8-44f7-92d0-65c2609dd30d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.067236] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-d54e0102-d52b-44c3-93dd-ba46528a03fc tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 938.067482] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-26efc9c7-3ac9-43f6-b365-8f126b58cf2b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.075364] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Acquiring lock "refresh_cache-87b25494-1853-4c7b-ae8b-6283992e895c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.075665] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Acquired lock "refresh_cache-87b25494-1853-4c7b-ae8b-6283992e895c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.075665] env[61868]: DEBUG nova.network.neutron [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 938.078618] env[61868]: DEBUG oslo_vmware.api [None req-d54e0102-d52b-44c3-93dd-ba46528a03fc tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 938.078618] env[61868]: value = "task-1315724" [ 938.078618] env[61868]: _type = "Task" [ 938.078618] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.088580] env[61868]: DEBUG oslo_vmware.api [None req-d54e0102-d52b-44c3-93dd-ba46528a03fc tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315724, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.103024] env[61868]: DEBUG nova.compute.manager [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 938.174688] env[61868]: DEBUG oslo_vmware.api [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Task: {'id': task-1315723, 'name': Rename_Task, 'duration_secs': 0.232532} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.175273] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 938.175381] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7fcf9825-f676-4b8e-87af-40fc6c12cf1e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.184663] env[61868]: DEBUG oslo_vmware.api [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Waiting for the task: (returnval){ [ 938.184663] env[61868]: value = "task-1315725" [ 938.184663] env[61868]: _type = "Task" [ 938.184663] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.194177] env[61868]: DEBUG oslo_vmware.api [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Task: {'id': task-1315725, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.238726] env[61868]: DEBUG oslo_concurrency.lockutils [None req-dfc04811-68b9-46c1-b4dd-5650499adb22 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "interface-5aec2609-79d3-4725-a182-70b575adbe17-87455df2-50af-4a63-b58b-4813a20d0b38" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.879s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.334938] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-766e00f4-77d1-4c6a-835b-00403fa16748 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.348100] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-952e1ba7-9360-4f45-a87f-ff87e9a09ee6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.357202] env[61868]: DEBUG oslo_vmware.api [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315722, 'name': ReconfigVM_Task, 'duration_secs': 0.547616} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.380715] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Updating instance '842c8606-632a-47d4-9a67-38d8e6ce6b4d' progress to 33 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 938.385071] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2233769-df81-470b-933c-e7725d299f37 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.394054] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d29c862f-d13e-4426-84b8-59404191d84e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.409068] env[61868]: DEBUG nova.compute.provider_tree [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 938.591488] env[61868]: DEBUG oslo_vmware.api [None req-d54e0102-d52b-44c3-93dd-ba46528a03fc tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315724, 'name': PowerOffVM_Task, 'duration_secs': 0.20156} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.592931] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-d54e0102-d52b-44c3-93dd-ba46528a03fc tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 938.593226] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-d54e0102-d52b-44c3-93dd-ba46528a03fc tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 938.593544] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-aecfe1e4-3d40-481b-84f0-80af93594c97 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.616691] env[61868]: DEBUG nova.network.neutron [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 938.684740] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-d54e0102-d52b-44c3-93dd-ba46528a03fc tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 938.685144] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-d54e0102-d52b-44c3-93dd-ba46528a03fc tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Deleting contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 938.685351] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-d54e0102-d52b-44c3-93dd-ba46528a03fc tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Deleting the datastore file [datastore2] 19c2720c-90bc-47f6-999b-6031f893408d {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 938.690629] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-629b2df5-997c-4420-93e2-6553c50b02ec {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.700103] env[61868]: DEBUG oslo_vmware.api [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Task: {'id': task-1315725, 'name': PowerOnVM_Task, 'duration_secs': 0.511072} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.703686] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 938.703888] env[61868]: INFO nova.compute.manager [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Took 6.69 seconds to spawn the instance on the hypervisor. [ 938.704101] env[61868]: DEBUG nova.compute.manager [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 938.704455] env[61868]: DEBUG oslo_vmware.api [None req-d54e0102-d52b-44c3-93dd-ba46528a03fc tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for the task: (returnval){ [ 938.704455] env[61868]: value = "task-1315727" [ 938.704455] env[61868]: _type = "Task" [ 938.704455] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.705317] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4d54fa3-3be3-49d6-9883-8965ee9a8d91 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.716963] env[61868]: DEBUG oslo_vmware.api [None req-d54e0102-d52b-44c3-93dd-ba46528a03fc tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315727, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.790341] env[61868]: DEBUG nova.network.neutron [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Updating instance_info_cache with network_info: [{"id": "bbcb967e-bf16-4589-ab6a-23e5ce9e8dd4", "address": "fa:16:3e:5e:7c:90", "network": {"id": "4545cca8-2f4d-4946-aa3c-2dbb36207e1f", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-229234284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0880f172b6534c92be8ba18e6cab17fe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbbcb967e-bf", "ovs_interfaceid": "bbcb967e-bf16-4589-ab6a-23e5ce9e8dd4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.890513] env[61868]: DEBUG nova.virt.hardware [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 938.890774] env[61868]: DEBUG nova.virt.hardware [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 938.890934] env[61868]: DEBUG nova.virt.hardware [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 938.891188] env[61868]: DEBUG nova.virt.hardware [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 938.891420] env[61868]: DEBUG nova.virt.hardware [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 938.891615] env[61868]: DEBUG nova.virt.hardware [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 938.891838] env[61868]: DEBUG nova.virt.hardware [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 938.892073] env[61868]: DEBUG nova.virt.hardware [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 938.892286] env[61868]: DEBUG nova.virt.hardware [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 938.892469] env[61868]: DEBUG nova.virt.hardware [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 938.892678] env[61868]: DEBUG nova.virt.hardware [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 938.897970] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Reconfiguring VM instance instance-0000004e to detach disk 2000 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 938.898279] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a34fb34a-baea-4764-9d6c-4dbaad0f7a96 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.911579] env[61868]: DEBUG nova.scheduler.client.report [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 938.921454] env[61868]: DEBUG oslo_vmware.api [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 938.921454] env[61868]: value = "task-1315728" [ 938.921454] env[61868]: _type = "Task" [ 938.921454] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.931766] env[61868]: DEBUG oslo_vmware.api [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315728, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.111698] env[61868]: DEBUG nova.compute.manager [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 939.133834] env[61868]: DEBUG nova.virt.hardware [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 939.134113] env[61868]: DEBUG nova.virt.hardware [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 939.134312] env[61868]: DEBUG nova.virt.hardware [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 939.134481] env[61868]: DEBUG nova.virt.hardware [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 939.134685] env[61868]: DEBUG nova.virt.hardware [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 939.134769] env[61868]: DEBUG nova.virt.hardware [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 939.134971] env[61868]: DEBUG nova.virt.hardware [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 939.135379] env[61868]: DEBUG nova.virt.hardware [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 939.135566] env[61868]: DEBUG nova.virt.hardware [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 939.135741] env[61868]: DEBUG nova.virt.hardware [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 939.135915] env[61868]: DEBUG nova.virt.hardware [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 939.136801] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc50edc1-e946-4bdc-a0d1-b334f2b9c1fc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.145954] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-478d60e2-572b-4850-9804-df782656e89b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.219136] env[61868]: DEBUG oslo_vmware.api [None req-d54e0102-d52b-44c3-93dd-ba46528a03fc tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Task: {'id': task-1315727, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.493731} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.219389] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-d54e0102-d52b-44c3-93dd-ba46528a03fc tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 939.219608] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-d54e0102-d52b-44c3-93dd-ba46528a03fc tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Deleted contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 939.219796] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-d54e0102-d52b-44c3-93dd-ba46528a03fc tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 939.219984] env[61868]: INFO nova.compute.manager [None req-d54e0102-d52b-44c3-93dd-ba46528a03fc tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Took 1.16 seconds to destroy the instance on the hypervisor. [ 939.220269] env[61868]: DEBUG oslo.service.loopingcall [None req-d54e0102-d52b-44c3-93dd-ba46528a03fc tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 939.220468] env[61868]: DEBUG nova.compute.manager [-] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 939.220559] env[61868]: DEBUG nova.network.neutron [-] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 939.228025] env[61868]: INFO nova.compute.manager [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Took 30.44 seconds to build instance. [ 939.293394] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Releasing lock "refresh_cache-87b25494-1853-4c7b-ae8b-6283992e895c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.293581] env[61868]: DEBUG nova.compute.manager [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Instance network_info: |[{"id": "bbcb967e-bf16-4589-ab6a-23e5ce9e8dd4", "address": "fa:16:3e:5e:7c:90", "network": {"id": "4545cca8-2f4d-4946-aa3c-2dbb36207e1f", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-229234284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0880f172b6534c92be8ba18e6cab17fe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbbcb967e-bf", "ovs_interfaceid": "bbcb967e-bf16-4589-ab6a-23e5ce9e8dd4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 939.294128] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5e:7c:90', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '089ef678-58b4-4bf0-a39d-b94b2d364291', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bbcb967e-bf16-4589-ab6a-23e5ce9e8dd4', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 939.302214] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Creating folder: Project (0880f172b6534c92be8ba18e6cab17fe). Parent ref: group-v281478. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 939.302429] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2a7048d2-e8a8-401f-b337-4ad31d5888d2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.316780] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Created folder: Project (0880f172b6534c92be8ba18e6cab17fe) in parent group-v281478. [ 939.317181] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Creating folder: Instances. Parent ref: group-v281585. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 939.317381] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-094fe0e5-98ba-4ffe-b1c5-8c2497699386 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.327641] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Created folder: Instances in parent group-v281585. [ 939.327942] env[61868]: DEBUG oslo.service.loopingcall [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 939.328204] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 939.328430] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4be60f50-b6b9-40aa-9130-bd25af343aa1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.350397] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 939.350397] env[61868]: value = "task-1315731" [ 939.350397] env[61868]: _type = "Task" [ 939.350397] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.360648] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315731, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.416616] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.312s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.417185] env[61868]: DEBUG nova.compute.manager [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 939.420236] env[61868]: DEBUG oslo_concurrency.lockutils [None req-40f8b9a6-0376-496a-97a7-807a4e018476 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.533s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.420486] env[61868]: DEBUG nova.objects.instance [None req-40f8b9a6-0376-496a-97a7-807a4e018476 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Lazy-loading 'resources' on Instance uuid 409cad1a-946d-4c58-aa57-1c0bf97fe63f {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 939.436757] env[61868]: DEBUG oslo_vmware.api [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315728, 'name': ReconfigVM_Task, 'duration_secs': 0.203829} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.437924] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Reconfigured VM instance instance-0000004e to detach disk 2000 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 939.439315] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87ea4b81-141e-4098-beff-e591197936d4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.476544] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] 842c8606-632a-47d4-9a67-38d8e6ce6b4d/842c8606-632a-47d4-9a67-38d8e6ce6b4d.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 939.477386] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3a0875e4-d19c-4ef1-b0d6-360bd6497189 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.497797] env[61868]: DEBUG oslo_vmware.api [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 939.497797] env[61868]: value = "task-1315732" [ 939.497797] env[61868]: _type = "Task" [ 939.497797] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.507095] env[61868]: DEBUG oslo_vmware.api [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315732, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.730942] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6530de9a-d200-4b61-aaef-1c60a93bb494 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Lock "6ce39c07-939c-49f3-8871-6e2163709f4a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.950s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.740830] env[61868]: DEBUG nova.compute.manager [req-fc011941-d8bc-4848-a93d-eabcee06a328 req-bbdf03e0-9fdf-4920-9f36-813f2d7171aa service nova] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Received event network-changed-bbcb967e-bf16-4589-ab6a-23e5ce9e8dd4 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 939.741035] env[61868]: DEBUG nova.compute.manager [req-fc011941-d8bc-4848-a93d-eabcee06a328 req-bbdf03e0-9fdf-4920-9f36-813f2d7171aa service nova] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Refreshing instance network info cache due to event network-changed-bbcb967e-bf16-4589-ab6a-23e5ce9e8dd4. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 939.741249] env[61868]: DEBUG oslo_concurrency.lockutils [req-fc011941-d8bc-4848-a93d-eabcee06a328 req-bbdf03e0-9fdf-4920-9f36-813f2d7171aa service nova] Acquiring lock "refresh_cache-87b25494-1853-4c7b-ae8b-6283992e895c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.741421] env[61868]: DEBUG oslo_concurrency.lockutils [req-fc011941-d8bc-4848-a93d-eabcee06a328 req-bbdf03e0-9fdf-4920-9f36-813f2d7171aa service nova] Acquired lock "refresh_cache-87b25494-1853-4c7b-ae8b-6283992e895c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.741546] env[61868]: DEBUG nova.network.neutron [req-fc011941-d8bc-4848-a93d-eabcee06a328 req-bbdf03e0-9fdf-4920-9f36-813f2d7171aa service nova] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Refreshing network info cache for port bbcb967e-bf16-4589-ab6a-23e5ce9e8dd4 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 939.811030] env[61868]: DEBUG nova.network.neutron [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Successfully updated port: 1f37fd44-6543-44b3-9fce-cb4154f0f0d0 {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 939.866786] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315731, 'name': CreateVM_Task, 'duration_secs': 0.415461} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.866989] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 939.867813] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.868487] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.868487] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 939.868692] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-94ef8135-43db-4091-bfee-1d1ce6cc42ad {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.878487] env[61868]: DEBUG oslo_vmware.api [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Waiting for the task: (returnval){ [ 939.878487] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]526dd37f-527d-9238-673b-5bab11781a73" [ 939.878487] env[61868]: _type = "Task" [ 939.878487] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.893033] env[61868]: DEBUG oslo_vmware.api [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]526dd37f-527d-9238-673b-5bab11781a73, 'name': SearchDatastore_Task} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.893033] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.893033] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 939.893033] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.893033] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.893033] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 939.893033] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-58e30570-5faf-4086-a30c-17a6916f662d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.905219] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 939.905219] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 939.905219] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0731705e-c1e5-4149-810b-d6acf6f2c8ab {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.914024] env[61868]: DEBUG oslo_vmware.api [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Waiting for the task: (returnval){ [ 939.914024] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5298c9b8-0351-f4f8-4828-cd25e2cdb80c" [ 939.914024] env[61868]: _type = "Task" [ 939.914024] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.925112] env[61868]: DEBUG nova.compute.utils [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 939.926246] env[61868]: DEBUG oslo_vmware.api [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5298c9b8-0351-f4f8-4828-cd25e2cdb80c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.926361] env[61868]: DEBUG nova.compute.manager [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 939.926558] env[61868]: DEBUG nova.network.neutron [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 939.977418] env[61868]: DEBUG nova.policy [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'daad105e7edf4fb0ae0b2e685bfd92e9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b520c3ac58074e8d9b0bfafb817244a8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 940.014583] env[61868]: DEBUG oslo_vmware.api [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315732, 'name': ReconfigVM_Task, 'duration_secs': 0.275274} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.014759] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Reconfigured VM instance instance-0000004e to attach disk [datastore2] 842c8606-632a-47d4-9a67-38d8e6ce6b4d/842c8606-632a-47d4-9a67-38d8e6ce6b4d.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 940.015058] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Updating instance '842c8606-632a-47d4-9a67-38d8e6ce6b4d' progress to 50 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 940.131252] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b811c2c-eed5-4f68-ad99-ee2163930427 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.139485] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bede8a7c-fbfe-4fdf-8508-f2ffc040006b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.173970] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06fa7cc6-4bfd-4fba-a31b-923e8c10f9e6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.182626] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-046b31b3-3140-413b-9e20-5acbab0b4540 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.200239] env[61868]: DEBUG nova.compute.provider_tree [None req-40f8b9a6-0376-496a-97a7-807a4e018476 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 940.313956] env[61868]: DEBUG oslo_concurrency.lockutils [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "refresh_cache-867384c6-175b-4848-b9ac-917470742ba2" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.314149] env[61868]: DEBUG oslo_concurrency.lockutils [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquired lock "refresh_cache-867384c6-175b-4848-b9ac-917470742ba2" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.314420] env[61868]: DEBUG nova.network.neutron [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 940.340806] env[61868]: DEBUG nova.network.neutron [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Successfully created port: 9c333eb1-c59f-4cf4-b25d-2f7a4c6c0a5e {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 940.358107] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0b461ce5-034e-41b9-8c34-be9526d11ff6 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Acquiring lock "6ce39c07-939c-49f3-8871-6e2163709f4a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.358107] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0b461ce5-034e-41b9-8c34-be9526d11ff6 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Lock "6ce39c07-939c-49f3-8871-6e2163709f4a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.358107] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0b461ce5-034e-41b9-8c34-be9526d11ff6 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Acquiring lock "6ce39c07-939c-49f3-8871-6e2163709f4a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.358107] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0b461ce5-034e-41b9-8c34-be9526d11ff6 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Lock "6ce39c07-939c-49f3-8871-6e2163709f4a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.358455] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0b461ce5-034e-41b9-8c34-be9526d11ff6 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Lock "6ce39c07-939c-49f3-8871-6e2163709f4a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.361239] env[61868]: INFO nova.compute.manager [None req-0b461ce5-034e-41b9-8c34-be9526d11ff6 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Terminating instance [ 940.368025] env[61868]: DEBUG nova.compute.manager [None req-0b461ce5-034e-41b9-8c34-be9526d11ff6 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 940.368639] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0b461ce5-034e-41b9-8c34-be9526d11ff6 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 940.370884] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2dff233-4337-4cd2-9819-a850c09a18a6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.378323] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b461ce5-034e-41b9-8c34-be9526d11ff6 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 940.378603] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-59aeb7f9-4186-4bae-9101-136d55a4f24c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.387979] env[61868]: DEBUG nova.network.neutron [-] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.387979] env[61868]: DEBUG oslo_vmware.api [None req-0b461ce5-034e-41b9-8c34-be9526d11ff6 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Waiting for the task: (returnval){ [ 940.387979] env[61868]: value = "task-1315733" [ 940.387979] env[61868]: _type = "Task" [ 940.387979] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.406975] env[61868]: DEBUG oslo_vmware.api [None req-0b461ce5-034e-41b9-8c34-be9526d11ff6 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Task: {'id': task-1315733, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.423204] env[61868]: DEBUG oslo_vmware.api [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5298c9b8-0351-f4f8-4828-cd25e2cdb80c, 'name': SearchDatastore_Task, 'duration_secs': 0.015402} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.424028] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff908128-ecdb-4201-ba61-c58602d5c9fd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.429478] env[61868]: DEBUG nova.compute.manager [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 940.436024] env[61868]: DEBUG oslo_vmware.api [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Waiting for the task: (returnval){ [ 940.436024] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]523e8f9f-bd3d-031b-5566-c5020cc1c0e4" [ 940.436024] env[61868]: _type = "Task" [ 940.436024] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.449595] env[61868]: DEBUG oslo_vmware.api [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]523e8f9f-bd3d-031b-5566-c5020cc1c0e4, 'name': SearchDatastore_Task, 'duration_secs': 0.009635} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.449595] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.449595] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 87b25494-1853-4c7b-ae8b-6283992e895c/87b25494-1853-4c7b-ae8b-6283992e895c.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 940.449595] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f4f609a0-b543-472c-a50b-48e08b3df171 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.458060] env[61868]: DEBUG oslo_vmware.api [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Waiting for the task: (returnval){ [ 940.458060] env[61868]: value = "task-1315734" [ 940.458060] env[61868]: _type = "Task" [ 940.458060] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.468850] env[61868]: DEBUG oslo_vmware.api [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Task: {'id': task-1315734, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.526052] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fe079bf-f926-48da-a33b-7358ca3cf29b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.550794] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-053e38b5-b537-485d-a5a4-2529402c83c3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.577914] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Updating instance '842c8606-632a-47d4-9a67-38d8e6ce6b4d' progress to 67 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 940.657102] env[61868]: DEBUG nova.network.neutron [req-fc011941-d8bc-4848-a93d-eabcee06a328 req-bbdf03e0-9fdf-4920-9f36-813f2d7171aa service nova] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Updated VIF entry in instance network info cache for port bbcb967e-bf16-4589-ab6a-23e5ce9e8dd4. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 940.657530] env[61868]: DEBUG nova.network.neutron [req-fc011941-d8bc-4848-a93d-eabcee06a328 req-bbdf03e0-9fdf-4920-9f36-813f2d7171aa service nova] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Updating instance_info_cache with network_info: [{"id": "bbcb967e-bf16-4589-ab6a-23e5ce9e8dd4", "address": "fa:16:3e:5e:7c:90", "network": {"id": "4545cca8-2f4d-4946-aa3c-2dbb36207e1f", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-229234284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0880f172b6534c92be8ba18e6cab17fe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbbcb967e-bf", "ovs_interfaceid": "bbcb967e-bf16-4589-ab6a-23e5ce9e8dd4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.704371] env[61868]: DEBUG nova.scheduler.client.report [None req-40f8b9a6-0376-496a-97a7-807a4e018476 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 940.713324] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d6dff523-f294-488f-8744-0eca0719bc34 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "interface-5aec2609-79d3-4725-a182-70b575adbe17-8a5f760c-c14e-4701-bde0-97fff3b6289e" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.713324] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d6dff523-f294-488f-8744-0eca0719bc34 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "interface-5aec2609-79d3-4725-a182-70b575adbe17-8a5f760c-c14e-4701-bde0-97fff3b6289e" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.856817] env[61868]: DEBUG nova.network.neutron [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 940.905193] env[61868]: INFO nova.compute.manager [-] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Took 1.68 seconds to deallocate network for instance. [ 940.905193] env[61868]: DEBUG oslo_vmware.api [None req-0b461ce5-034e-41b9-8c34-be9526d11ff6 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Task: {'id': task-1315733, 'name': PowerOffVM_Task, 'duration_secs': 0.216119} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.906707] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b461ce5-034e-41b9-8c34-be9526d11ff6 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 940.907160] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0b461ce5-034e-41b9-8c34-be9526d11ff6 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 940.907569] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d078b99e-9b1d-4847-8bfe-471489abee21 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.973363] env[61868]: DEBUG oslo_vmware.api [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Task: {'id': task-1315734, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.509591} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.975101] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 87b25494-1853-4c7b-ae8b-6283992e895c/87b25494-1853-4c7b-ae8b-6283992e895c.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 940.975101] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 940.975331] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2c86131d-ce7c-40fa-8cc2-c308cf9de49e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.985770] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0b461ce5-034e-41b9-8c34-be9526d11ff6 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 940.988016] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0b461ce5-034e-41b9-8c34-be9526d11ff6 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 940.988016] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b461ce5-034e-41b9-8c34-be9526d11ff6 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Deleting the datastore file [datastore1] 6ce39c07-939c-49f3-8871-6e2163709f4a {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 940.988540] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7be00b9c-eb7c-4d71-81a6-80aae8c7cb70 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.991281] env[61868]: DEBUG oslo_vmware.api [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Waiting for the task: (returnval){ [ 940.991281] env[61868]: value = "task-1315736" [ 940.991281] env[61868]: _type = "Task" [ 940.991281] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.997488] env[61868]: DEBUG oslo_vmware.api [None req-0b461ce5-034e-41b9-8c34-be9526d11ff6 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Waiting for the task: (returnval){ [ 940.997488] env[61868]: value = "task-1315737" [ 940.997488] env[61868]: _type = "Task" [ 940.997488] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.009989] env[61868]: DEBUG oslo_vmware.api [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Task: {'id': task-1315736, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.015625] env[61868]: DEBUG oslo_vmware.api [None req-0b461ce5-034e-41b9-8c34-be9526d11ff6 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Task: {'id': task-1315737, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.052350] env[61868]: DEBUG nova.network.neutron [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Updating instance_info_cache with network_info: [{"id": "1f37fd44-6543-44b3-9fce-cb4154f0f0d0", "address": "fa:16:3e:22:35:19", "network": {"id": "f2e3f911-d826-43ad-911d-9f6985db7ee5", "bridge": "br-int", "label": "tempest-ImagesTestJSON-485234144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5658848e1b0e42929c04a4a8de40a291", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f37fd44-65", "ovs_interfaceid": "1f37fd44-6543-44b3-9fce-cb4154f0f0d0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.122916] env[61868]: DEBUG nova.network.neutron [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Port e3943fa9-423d-4168-8aa6-1651280cd127 binding to destination host cpu-1 is already ACTIVE {{(pid=61868) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 941.160823] env[61868]: DEBUG oslo_concurrency.lockutils [req-fc011941-d8bc-4848-a93d-eabcee06a328 req-bbdf03e0-9fdf-4920-9f36-813f2d7171aa service nova] Releasing lock "refresh_cache-87b25494-1853-4c7b-ae8b-6283992e895c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.210347] env[61868]: DEBUG oslo_concurrency.lockutils [None req-40f8b9a6-0376-496a-97a7-807a4e018476 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.790s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.212720] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 8.500s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.213369] env[61868]: DEBUG nova.objects.instance [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61868) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 941.217121] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d6dff523-f294-488f-8744-0eca0719bc34 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.217305] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d6dff523-f294-488f-8744-0eca0719bc34 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquired lock "5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.218483] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9521e017-ef5b-4ef5-a9b1-5a97626a092e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.240824] env[61868]: INFO nova.scheduler.client.report [None req-40f8b9a6-0376-496a-97a7-807a4e018476 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Deleted allocations for instance 409cad1a-946d-4c58-aa57-1c0bf97fe63f [ 941.244796] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72b44cef-8af2-4d98-a918-cb5c5dc9fa5f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.281267] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-d6dff523-f294-488f-8744-0eca0719bc34 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Reconfiguring VM to detach interface {{(pid=61868) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 941.281913] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-acc2f548-6515-46f0-9c08-9557cc803329 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.303063] env[61868]: DEBUG oslo_vmware.api [None req-d6dff523-f294-488f-8744-0eca0719bc34 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for the task: (returnval){ [ 941.303063] env[61868]: value = "task-1315738" [ 941.303063] env[61868]: _type = "Task" [ 941.303063] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.313509] env[61868]: DEBUG oslo_vmware.api [None req-d6dff523-f294-488f-8744-0eca0719bc34 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315738, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.416085] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d54e0102-d52b-44c3-93dd-ba46528a03fc tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.442068] env[61868]: DEBUG nova.compute.manager [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 941.468649] env[61868]: DEBUG nova.virt.hardware [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 941.468911] env[61868]: DEBUG nova.virt.hardware [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 941.469099] env[61868]: DEBUG nova.virt.hardware [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 941.469295] env[61868]: DEBUG nova.virt.hardware [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 941.469464] env[61868]: DEBUG nova.virt.hardware [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 941.469702] env[61868]: DEBUG nova.virt.hardware [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 941.469944] env[61868]: DEBUG nova.virt.hardware [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 941.470135] env[61868]: DEBUG nova.virt.hardware [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 941.470313] env[61868]: DEBUG nova.virt.hardware [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 941.470485] env[61868]: DEBUG nova.virt.hardware [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 941.470665] env[61868]: DEBUG nova.virt.hardware [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 941.471642] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1efa078e-fe69-4639-92cf-7bcb8109cd4e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.480639] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee6ccad7-9407-4d5c-bdee-517214a536cc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.514811] env[61868]: DEBUG oslo_vmware.api [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Task: {'id': task-1315736, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064556} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.518736] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 941.519148] env[61868]: DEBUG oslo_vmware.api [None req-0b461ce5-034e-41b9-8c34-be9526d11ff6 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Task: {'id': task-1315737, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14556} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.520533] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbadfb80-7424-4106-b806-c24a0b00ad0f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.523169] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b461ce5-034e-41b9-8c34-be9526d11ff6 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 941.523386] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0b461ce5-034e-41b9-8c34-be9526d11ff6 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 941.523569] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0b461ce5-034e-41b9-8c34-be9526d11ff6 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 941.523751] env[61868]: INFO nova.compute.manager [None req-0b461ce5-034e-41b9-8c34-be9526d11ff6 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Took 1.16 seconds to destroy the instance on the hypervisor. [ 941.523994] env[61868]: DEBUG oslo.service.loopingcall [None req-0b461ce5-034e-41b9-8c34-be9526d11ff6 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 941.524861] env[61868]: DEBUG nova.compute.manager [-] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 941.524861] env[61868]: DEBUG nova.network.neutron [-] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 941.548140] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] 87b25494-1853-4c7b-ae8b-6283992e895c/87b25494-1853-4c7b-ae8b-6283992e895c.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 941.548550] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-16a39a4a-ef98-4a1f-b34f-3984ea88de3e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.564275] env[61868]: DEBUG oslo_concurrency.lockutils [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Releasing lock "refresh_cache-867384c6-175b-4848-b9ac-917470742ba2" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.564623] env[61868]: DEBUG nova.compute.manager [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Instance network_info: |[{"id": "1f37fd44-6543-44b3-9fce-cb4154f0f0d0", "address": "fa:16:3e:22:35:19", "network": {"id": "f2e3f911-d826-43ad-911d-9f6985db7ee5", "bridge": "br-int", "label": "tempest-ImagesTestJSON-485234144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5658848e1b0e42929c04a4a8de40a291", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f37fd44-65", "ovs_interfaceid": "1f37fd44-6543-44b3-9fce-cb4154f0f0d0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 941.565433] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:22:35:19', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f1b507ed-cd2d-4c09-9d96-c47bde6a7774', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1f37fd44-6543-44b3-9fce-cb4154f0f0d0', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 941.572710] env[61868]: DEBUG oslo.service.loopingcall [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 941.574339] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 941.574627] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3685f77a-43ad-429e-a8eb-7f6a13295673 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.590991] env[61868]: DEBUG oslo_vmware.api [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Waiting for the task: (returnval){ [ 941.590991] env[61868]: value = "task-1315739" [ 941.590991] env[61868]: _type = "Task" [ 941.590991] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.598026] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 941.598026] env[61868]: value = "task-1315740" [ 941.598026] env[61868]: _type = "Task" [ 941.598026] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.601727] env[61868]: DEBUG oslo_vmware.api [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Task: {'id': task-1315739, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.613236] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315740, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.753963] env[61868]: DEBUG oslo_concurrency.lockutils [None req-40f8b9a6-0376-496a-97a7-807a4e018476 tempest-AttachVolumeTestJSON-997304659 tempest-AttachVolumeTestJSON-997304659-project-member] Lock "409cad1a-946d-4c58-aa57-1c0bf97fe63f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.265s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.774971] env[61868]: DEBUG nova.compute.manager [req-6c3e86f8-6c25-44a3-8144-b8693db675f6 req-62886dff-8a00-4bf0-8ad8-5f00e14f93bd service nova] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Received event network-vif-plugged-1f37fd44-6543-44b3-9fce-cb4154f0f0d0 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 941.775345] env[61868]: DEBUG oslo_concurrency.lockutils [req-6c3e86f8-6c25-44a3-8144-b8693db675f6 req-62886dff-8a00-4bf0-8ad8-5f00e14f93bd service nova] Acquiring lock "867384c6-175b-4848-b9ac-917470742ba2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.775759] env[61868]: DEBUG oslo_concurrency.lockutils [req-6c3e86f8-6c25-44a3-8144-b8693db675f6 req-62886dff-8a00-4bf0-8ad8-5f00e14f93bd service nova] Lock "867384c6-175b-4848-b9ac-917470742ba2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.776047] env[61868]: DEBUG oslo_concurrency.lockutils [req-6c3e86f8-6c25-44a3-8144-b8693db675f6 req-62886dff-8a00-4bf0-8ad8-5f00e14f93bd service nova] Lock "867384c6-175b-4848-b9ac-917470742ba2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.776346] env[61868]: DEBUG nova.compute.manager [req-6c3e86f8-6c25-44a3-8144-b8693db675f6 req-62886dff-8a00-4bf0-8ad8-5f00e14f93bd service nova] [instance: 867384c6-175b-4848-b9ac-917470742ba2] No waiting events found dispatching network-vif-plugged-1f37fd44-6543-44b3-9fce-cb4154f0f0d0 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 941.776651] env[61868]: WARNING nova.compute.manager [req-6c3e86f8-6c25-44a3-8144-b8693db675f6 req-62886dff-8a00-4bf0-8ad8-5f00e14f93bd service nova] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Received unexpected event network-vif-plugged-1f37fd44-6543-44b3-9fce-cb4154f0f0d0 for instance with vm_state building and task_state spawning. [ 941.777083] env[61868]: DEBUG nova.compute.manager [req-6c3e86f8-6c25-44a3-8144-b8693db675f6 req-62886dff-8a00-4bf0-8ad8-5f00e14f93bd service nova] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Received event network-changed-1f37fd44-6543-44b3-9fce-cb4154f0f0d0 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 941.777083] env[61868]: DEBUG nova.compute.manager [req-6c3e86f8-6c25-44a3-8144-b8693db675f6 req-62886dff-8a00-4bf0-8ad8-5f00e14f93bd service nova] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Refreshing instance network info cache due to event network-changed-1f37fd44-6543-44b3-9fce-cb4154f0f0d0. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 941.777240] env[61868]: DEBUG oslo_concurrency.lockutils [req-6c3e86f8-6c25-44a3-8144-b8693db675f6 req-62886dff-8a00-4bf0-8ad8-5f00e14f93bd service nova] Acquiring lock "refresh_cache-867384c6-175b-4848-b9ac-917470742ba2" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.777385] env[61868]: DEBUG oslo_concurrency.lockutils [req-6c3e86f8-6c25-44a3-8144-b8693db675f6 req-62886dff-8a00-4bf0-8ad8-5f00e14f93bd service nova] Acquired lock "refresh_cache-867384c6-175b-4848-b9ac-917470742ba2" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.777549] env[61868]: DEBUG nova.network.neutron [req-6c3e86f8-6c25-44a3-8144-b8693db675f6 req-62886dff-8a00-4bf0-8ad8-5f00e14f93bd service nova] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Refreshing network info cache for port 1f37fd44-6543-44b3-9fce-cb4154f0f0d0 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 941.814966] env[61868]: DEBUG oslo_vmware.api [None req-d6dff523-f294-488f-8744-0eca0719bc34 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315738, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.012088] env[61868]: DEBUG oslo_concurrency.lockutils [None req-294190a7-023a-4872-9a21-68724531565d tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Acquiring lock "4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.012088] env[61868]: DEBUG oslo_concurrency.lockutils [None req-294190a7-023a-4872-9a21-68724531565d tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Lock "4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.012088] env[61868]: DEBUG oslo_concurrency.lockutils [None req-294190a7-023a-4872-9a21-68724531565d tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Acquiring lock "4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.012088] env[61868]: DEBUG oslo_concurrency.lockutils [None req-294190a7-023a-4872-9a21-68724531565d tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Lock "4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.012088] env[61868]: DEBUG oslo_concurrency.lockutils [None req-294190a7-023a-4872-9a21-68724531565d tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Lock "4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.013046] env[61868]: INFO nova.compute.manager [None req-294190a7-023a-4872-9a21-68724531565d tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Terminating instance [ 942.015240] env[61868]: DEBUG nova.compute.manager [None req-294190a7-023a-4872-9a21-68724531565d tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 942.015604] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-294190a7-023a-4872-9a21-68724531565d tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 942.016583] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a7c60da-6c3f-42da-baa1-5cf84847cd96 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.027189] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-294190a7-023a-4872-9a21-68724531565d tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 942.027189] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6c6a8f2a-2af4-4e2f-aea9-f9158c2aa808 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.035512] env[61868]: DEBUG oslo_vmware.api [None req-294190a7-023a-4872-9a21-68724531565d tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Waiting for the task: (returnval){ [ 942.035512] env[61868]: value = "task-1315741" [ 942.035512] env[61868]: _type = "Task" [ 942.035512] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.047665] env[61868]: DEBUG oslo_vmware.api [None req-294190a7-023a-4872-9a21-68724531565d tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Task: {'id': task-1315741, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.052691] env[61868]: DEBUG nova.network.neutron [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Successfully updated port: 9c333eb1-c59f-4cf4-b25d-2f7a4c6c0a5e {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 942.103716] env[61868]: DEBUG oslo_vmware.api [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Task: {'id': task-1315739, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.113438] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315740, 'name': CreateVM_Task} progress is 25%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.151698] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "842c8606-632a-47d4-9a67-38d8e6ce6b4d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.151951] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "842c8606-632a-47d4-9a67-38d8e6ce6b4d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.152210] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "842c8606-632a-47d4-9a67-38d8e6ce6b4d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.223728] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c65bd449-9c62-42ed-bdbb-21fd24e9d5f6 tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.226082] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7f844117-d3b3-43c3-86a4-c26b82f8c10c tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.125s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.227575] env[61868]: DEBUG nova.objects.instance [None req-7f844117-d3b3-43c3-86a4-c26b82f8c10c tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lazy-loading 'resources' on Instance uuid 4e9d75e8-ca33-4e33-baa3-cb7d596f584e {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 942.315947] env[61868]: DEBUG oslo_vmware.api [None req-d6dff523-f294-488f-8744-0eca0719bc34 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315738, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.332404] env[61868]: DEBUG nova.network.neutron [-] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.520613] env[61868]: DEBUG nova.network.neutron [req-6c3e86f8-6c25-44a3-8144-b8693db675f6 req-62886dff-8a00-4bf0-8ad8-5f00e14f93bd service nova] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Updated VIF entry in instance network info cache for port 1f37fd44-6543-44b3-9fce-cb4154f0f0d0. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 942.520613] env[61868]: DEBUG nova.network.neutron [req-6c3e86f8-6c25-44a3-8144-b8693db675f6 req-62886dff-8a00-4bf0-8ad8-5f00e14f93bd service nova] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Updating instance_info_cache with network_info: [{"id": "1f37fd44-6543-44b3-9fce-cb4154f0f0d0", "address": "fa:16:3e:22:35:19", "network": {"id": "f2e3f911-d826-43ad-911d-9f6985db7ee5", "bridge": "br-int", "label": "tempest-ImagesTestJSON-485234144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5658848e1b0e42929c04a4a8de40a291", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f37fd44-65", "ovs_interfaceid": "1f37fd44-6543-44b3-9fce-cb4154f0f0d0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.550956] env[61868]: DEBUG oslo_vmware.api [None req-294190a7-023a-4872-9a21-68724531565d tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Task: {'id': task-1315741, 'name': PowerOffVM_Task, 'duration_secs': 0.264119} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.550956] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-294190a7-023a-4872-9a21-68724531565d tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 942.550956] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-294190a7-023a-4872-9a21-68724531565d tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 942.550956] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-341cd241-c986-46a9-986f-c47fe674e63e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.557485] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "refresh_cache-cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.557485] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquired lock "refresh_cache-cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.557485] env[61868]: DEBUG nova.network.neutron [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 942.606027] env[61868]: DEBUG oslo_vmware.api [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Task: {'id': task-1315739, 'name': ReconfigVM_Task, 'duration_secs': 0.711052} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.606027] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Reconfigured VM instance instance-00000054 to attach disk [datastore1] 87b25494-1853-4c7b-ae8b-6283992e895c/87b25494-1853-4c7b-ae8b-6283992e895c.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 942.608170] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-665b93b4-1533-48f9-b15a-7a27724605fe {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.614460] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-294190a7-023a-4872-9a21-68724531565d tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 942.614922] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-294190a7-023a-4872-9a21-68724531565d tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 942.615245] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-294190a7-023a-4872-9a21-68724531565d tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Deleting the datastore file [datastore1] 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 942.619898] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9e357995-8ec2-4716-8ea3-6b0e8725d160 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.622429] env[61868]: DEBUG oslo_vmware.api [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Waiting for the task: (returnval){ [ 942.622429] env[61868]: value = "task-1315743" [ 942.622429] env[61868]: _type = "Task" [ 942.622429] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.622874] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315740, 'name': CreateVM_Task, 'duration_secs': 0.632954} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.623260] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 942.627462] env[61868]: DEBUG oslo_concurrency.lockutils [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.627842] env[61868]: DEBUG oslo_concurrency.lockutils [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.628277] env[61868]: DEBUG oslo_concurrency.lockutils [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 942.629718] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf0c8dce-7e85-4b09-96ec-eb35322b6fde {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.632114] env[61868]: DEBUG oslo_vmware.api [None req-294190a7-023a-4872-9a21-68724531565d tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Waiting for the task: (returnval){ [ 942.632114] env[61868]: value = "task-1315744" [ 942.632114] env[61868]: _type = "Task" [ 942.632114] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.640182] env[61868]: DEBUG oslo_vmware.api [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Task: {'id': task-1315743, 'name': Rename_Task} progress is 10%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.640634] env[61868]: DEBUG oslo_vmware.api [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 942.640634] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52066600-e117-72b4-5370-a31c955df863" [ 942.640634] env[61868]: _type = "Task" [ 942.640634] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.650495] env[61868]: DEBUG oslo_vmware.api [None req-294190a7-023a-4872-9a21-68724531565d tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Task: {'id': task-1315744, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.659744] env[61868]: DEBUG oslo_vmware.api [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52066600-e117-72b4-5370-a31c955df863, 'name': SearchDatastore_Task, 'duration_secs': 0.009678} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.659744] env[61868]: DEBUG oslo_concurrency.lockutils [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.659744] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 942.659960] env[61868]: DEBUG oslo_concurrency.lockutils [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.660996] env[61868]: DEBUG oslo_concurrency.lockutils [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.660996] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 942.660996] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1f61d268-2906-41b7-a248-8c0966e1394f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.670334] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 942.670533] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 942.671541] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5f6d5611-15ee-4c0d-86b9-a8821ab2a68d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.678316] env[61868]: DEBUG oslo_vmware.api [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 942.678316] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52ad2baf-753e-caba-30c3-c593c0216ab7" [ 942.678316] env[61868]: _type = "Task" [ 942.678316] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.687849] env[61868]: DEBUG oslo_vmware.api [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52ad2baf-753e-caba-30c3-c593c0216ab7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.818190] env[61868]: DEBUG oslo_vmware.api [None req-d6dff523-f294-488f-8744-0eca0719bc34 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315738, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.834551] env[61868]: INFO nova.compute.manager [-] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Took 1.31 seconds to deallocate network for instance. [ 942.958884] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa08f411-a5a0-46bb-b8bc-0aab64bad937 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.967077] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34a91b27-db70-45b4-8f30-357cf0bc9712 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.997838] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38e258bb-7486-4968-a2c0-c26e881f7f48 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.007267] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db88a438-d207-4bd7-9772-9b1a4910346b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.021305] env[61868]: DEBUG nova.compute.provider_tree [None req-7f844117-d3b3-43c3-86a4-c26b82f8c10c tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 943.022874] env[61868]: DEBUG oslo_concurrency.lockutils [req-6c3e86f8-6c25-44a3-8144-b8693db675f6 req-62886dff-8a00-4bf0-8ad8-5f00e14f93bd service nova] Releasing lock "refresh_cache-867384c6-175b-4848-b9ac-917470742ba2" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.023151] env[61868]: DEBUG nova.compute.manager [req-6c3e86f8-6c25-44a3-8144-b8693db675f6 req-62886dff-8a00-4bf0-8ad8-5f00e14f93bd service nova] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Received event network-vif-deleted-77081d78-be71-47a0-a7be-8511ae0243d4 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 943.087777] env[61868]: DEBUG nova.network.neutron [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 943.136168] env[61868]: DEBUG oslo_vmware.api [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Task: {'id': task-1315743, 'name': Rename_Task, 'duration_secs': 0.149324} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.139167] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 943.139430] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-43bb8997-afca-451c-92b4-2a95196bc06c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.148768] env[61868]: DEBUG oslo_vmware.api [None req-294190a7-023a-4872-9a21-68724531565d tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Task: {'id': task-1315744, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.142786} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.150042] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-294190a7-023a-4872-9a21-68724531565d tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 943.150330] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-294190a7-023a-4872-9a21-68724531565d tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 943.150510] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-294190a7-023a-4872-9a21-68724531565d tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 943.150652] env[61868]: INFO nova.compute.manager [None req-294190a7-023a-4872-9a21-68724531565d tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Took 1.14 seconds to destroy the instance on the hypervisor. [ 943.150892] env[61868]: DEBUG oslo.service.loopingcall [None req-294190a7-023a-4872-9a21-68724531565d tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 943.151106] env[61868]: DEBUG oslo_vmware.api [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Waiting for the task: (returnval){ [ 943.151106] env[61868]: value = "task-1315746" [ 943.151106] env[61868]: _type = "Task" [ 943.151106] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.151309] env[61868]: DEBUG nova.compute.manager [-] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 943.151402] env[61868]: DEBUG nova.network.neutron [-] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 943.165398] env[61868]: DEBUG oslo_vmware.api [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Task: {'id': task-1315746, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.190076] env[61868]: DEBUG oslo_vmware.api [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52ad2baf-753e-caba-30c3-c593c0216ab7, 'name': SearchDatastore_Task, 'duration_secs': 0.009612} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.190953] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d9bc58a3-dc0a-4d93-a1cc-136640f3a6e5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.198856] env[61868]: DEBUG oslo_vmware.api [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 943.198856] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5254c53e-9588-1be3-c448-0278f69e4414" [ 943.198856] env[61868]: _type = "Task" [ 943.198856] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.207605] env[61868]: DEBUG oslo_vmware.api [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5254c53e-9588-1be3-c448-0278f69e4414, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.246249] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "refresh_cache-842c8606-632a-47d4-9a67-38d8e6ce6b4d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.246596] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquired lock "refresh_cache-842c8606-632a-47d4-9a67-38d8e6ce6b4d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.246642] env[61868]: DEBUG nova.network.neutron [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 943.306123] env[61868]: DEBUG nova.network.neutron [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Updating instance_info_cache with network_info: [{"id": "9c333eb1-c59f-4cf4-b25d-2f7a4c6c0a5e", "address": "fa:16:3e:fa:c4:06", "network": {"id": "77f33b52-e4eb-4b72-85d4-5991a2b5c36d", "bridge": "br-int", "label": "tempest-ServersTestJSON-748635995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b520c3ac58074e8d9b0bfafb817244a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c333eb1-c5", "ovs_interfaceid": "9c333eb1-c59f-4cf4-b25d-2f7a4c6c0a5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.319208] env[61868]: DEBUG oslo_vmware.api [None req-d6dff523-f294-488f-8744-0eca0719bc34 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315738, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.341139] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0b461ce5-034e-41b9-8c34-be9526d11ff6 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.525140] env[61868]: DEBUG nova.scheduler.client.report [None req-7f844117-d3b3-43c3-86a4-c26b82f8c10c tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 943.664260] env[61868]: DEBUG oslo_vmware.api [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Task: {'id': task-1315746, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.711323] env[61868]: DEBUG oslo_vmware.api [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5254c53e-9588-1be3-c448-0278f69e4414, 'name': SearchDatastore_Task, 'duration_secs': 0.010835} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.711323] env[61868]: DEBUG oslo_concurrency.lockutils [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.711323] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 867384c6-175b-4848-b9ac-917470742ba2/867384c6-175b-4848-b9ac-917470742ba2.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 943.711323] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4db98227-0578-44a9-a59f-2fd67f9a9146 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.719171] env[61868]: DEBUG oslo_vmware.api [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 943.719171] env[61868]: value = "task-1315747" [ 943.719171] env[61868]: _type = "Task" [ 943.719171] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.728429] env[61868]: DEBUG oslo_vmware.api [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315747, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.814517] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Releasing lock "refresh_cache-cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.814577] env[61868]: DEBUG nova.compute.manager [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Instance network_info: |[{"id": "9c333eb1-c59f-4cf4-b25d-2f7a4c6c0a5e", "address": "fa:16:3e:fa:c4:06", "network": {"id": "77f33b52-e4eb-4b72-85d4-5991a2b5c36d", "bridge": "br-int", "label": "tempest-ServersTestJSON-748635995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b520c3ac58074e8d9b0bfafb817244a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c333eb1-c5", "ovs_interfaceid": "9c333eb1-c59f-4cf4-b25d-2f7a4c6c0a5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 943.815469] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fa:c4:06', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56136ef6-99d7-4562-9a9f-d66fec951c5c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9c333eb1-c59f-4cf4-b25d-2f7a4c6c0a5e', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 943.824262] env[61868]: DEBUG oslo.service.loopingcall [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 943.830446] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 943.832835] env[61868]: DEBUG nova.compute.manager [req-6ba70ff4-8394-4d7c-9f31-3bad47a6af19 req-e17a05df-dd20-4e7b-84e2-6d6160fcf461 service nova] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Received event network-vif-deleted-5ef47555-42a4-44dc-bf5f-01d21d921b5f {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 943.833192] env[61868]: DEBUG nova.compute.manager [req-6ba70ff4-8394-4d7c-9f31-3bad47a6af19 req-e17a05df-dd20-4e7b-84e2-6d6160fcf461 service nova] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Received event network-vif-plugged-9c333eb1-c59f-4cf4-b25d-2f7a4c6c0a5e {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 943.833468] env[61868]: DEBUG oslo_concurrency.lockutils [req-6ba70ff4-8394-4d7c-9f31-3bad47a6af19 req-e17a05df-dd20-4e7b-84e2-6d6160fcf461 service nova] Acquiring lock "cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.833763] env[61868]: DEBUG oslo_concurrency.lockutils [req-6ba70ff4-8394-4d7c-9f31-3bad47a6af19 req-e17a05df-dd20-4e7b-84e2-6d6160fcf461 service nova] Lock "cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.834009] env[61868]: DEBUG oslo_concurrency.lockutils [req-6ba70ff4-8394-4d7c-9f31-3bad47a6af19 req-e17a05df-dd20-4e7b-84e2-6d6160fcf461 service nova] Lock "cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.834248] env[61868]: DEBUG nova.compute.manager [req-6ba70ff4-8394-4d7c-9f31-3bad47a6af19 req-e17a05df-dd20-4e7b-84e2-6d6160fcf461 service nova] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] No waiting events found dispatching network-vif-plugged-9c333eb1-c59f-4cf4-b25d-2f7a4c6c0a5e {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 943.834508] env[61868]: WARNING nova.compute.manager [req-6ba70ff4-8394-4d7c-9f31-3bad47a6af19 req-e17a05df-dd20-4e7b-84e2-6d6160fcf461 service nova] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Received unexpected event network-vif-plugged-9c333eb1-c59f-4cf4-b25d-2f7a4c6c0a5e for instance with vm_state building and task_state spawning. [ 943.834808] env[61868]: DEBUG nova.compute.manager [req-6ba70ff4-8394-4d7c-9f31-3bad47a6af19 req-e17a05df-dd20-4e7b-84e2-6d6160fcf461 service nova] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Received event network-changed-9c333eb1-c59f-4cf4-b25d-2f7a4c6c0a5e {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 943.835104] env[61868]: DEBUG nova.compute.manager [req-6ba70ff4-8394-4d7c-9f31-3bad47a6af19 req-e17a05df-dd20-4e7b-84e2-6d6160fcf461 service nova] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Refreshing instance network info cache due to event network-changed-9c333eb1-c59f-4cf4-b25d-2f7a4c6c0a5e. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 943.835415] env[61868]: DEBUG oslo_concurrency.lockutils [req-6ba70ff4-8394-4d7c-9f31-3bad47a6af19 req-e17a05df-dd20-4e7b-84e2-6d6160fcf461 service nova] Acquiring lock "refresh_cache-cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.835677] env[61868]: DEBUG oslo_concurrency.lockutils [req-6ba70ff4-8394-4d7c-9f31-3bad47a6af19 req-e17a05df-dd20-4e7b-84e2-6d6160fcf461 service nova] Acquired lock "refresh_cache-cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.835960] env[61868]: DEBUG nova.network.neutron [req-6ba70ff4-8394-4d7c-9f31-3bad47a6af19 req-e17a05df-dd20-4e7b-84e2-6d6160fcf461 service nova] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Refreshing network info cache for port 9c333eb1-c59f-4cf4-b25d-2f7a4c6c0a5e {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 943.838044] env[61868]: DEBUG oslo_vmware.api [None req-d6dff523-f294-488f-8744-0eca0719bc34 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315738, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.841206] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5a5bffa7-0b86-4bc3-9add-6dd039e971a6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.868935] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 943.868935] env[61868]: value = "task-1315748" [ 943.868935] env[61868]: _type = "Task" [ 943.868935] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.879842] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315748, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.985406] env[61868]: DEBUG nova.network.neutron [-] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.030790] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7f844117-d3b3-43c3-86a4-c26b82f8c10c tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.804s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.038238] env[61868]: DEBUG oslo_concurrency.lockutils [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.912s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.038238] env[61868]: DEBUG nova.objects.instance [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Lazy-loading 'resources' on Instance uuid b51e968e-6911-47f3-b5a9-e8a5865e42a0 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 944.062926] env[61868]: INFO nova.scheduler.client.report [None req-7f844117-d3b3-43c3-86a4-c26b82f8c10c tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Deleted allocations for instance 4e9d75e8-ca33-4e33-baa3-cb7d596f584e [ 944.160080] env[61868]: DEBUG nova.network.neutron [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Updating instance_info_cache with network_info: [{"id": "e3943fa9-423d-4168-8aa6-1651280cd127", "address": "fa:16:3e:1d:68:a3", "network": {"id": "c1f5bf03-438f-4acc-93fd-159bed7d4069", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2061742146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a60c3c20950e4156b64c3b4c61b9f0f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24b91e3b-50f1-4a16-b929-942e6b31b2d8", "external-id": "nsx-vlan-transportzone-34", "segmentation_id": 34, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3943fa9-42", "ovs_interfaceid": "e3943fa9-423d-4168-8aa6-1651280cd127", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.165590] env[61868]: DEBUG oslo_vmware.api [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Task: {'id': task-1315746, 'name': PowerOnVM_Task, 'duration_secs': 0.728202} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.165935] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 944.166196] env[61868]: INFO nova.compute.manager [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Took 7.44 seconds to spawn the instance on the hypervisor. [ 944.166426] env[61868]: DEBUG nova.compute.manager [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 944.167318] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba302615-e2ae-4cc9-b23b-fd0dcc6a6cb9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.225724] env[61868]: DEBUG nova.network.neutron [req-6ba70ff4-8394-4d7c-9f31-3bad47a6af19 req-e17a05df-dd20-4e7b-84e2-6d6160fcf461 service nova] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Updated VIF entry in instance network info cache for port 9c333eb1-c59f-4cf4-b25d-2f7a4c6c0a5e. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 944.226267] env[61868]: DEBUG nova.network.neutron [req-6ba70ff4-8394-4d7c-9f31-3bad47a6af19 req-e17a05df-dd20-4e7b-84e2-6d6160fcf461 service nova] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Updating instance_info_cache with network_info: [{"id": "9c333eb1-c59f-4cf4-b25d-2f7a4c6c0a5e", "address": "fa:16:3e:fa:c4:06", "network": {"id": "77f33b52-e4eb-4b72-85d4-5991a2b5c36d", "bridge": "br-int", "label": "tempest-ServersTestJSON-748635995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b520c3ac58074e8d9b0bfafb817244a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c333eb1-c5", "ovs_interfaceid": "9c333eb1-c59f-4cf4-b25d-2f7a4c6c0a5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.235840] env[61868]: DEBUG oslo_vmware.api [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315747, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.319348] env[61868]: DEBUG oslo_vmware.api [None req-d6dff523-f294-488f-8744-0eca0719bc34 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315738, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.379123] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315748, 'name': CreateVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.490530] env[61868]: INFO nova.compute.manager [-] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Took 1.34 seconds to deallocate network for instance. [ 944.579561] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7f844117-d3b3-43c3-86a4-c26b82f8c10c tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "4e9d75e8-ca33-4e33-baa3-cb7d596f584e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.887s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.662936] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Releasing lock "refresh_cache-842c8606-632a-47d4-9a67-38d8e6ce6b4d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.691074] env[61868]: INFO nova.compute.manager [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Took 27.85 seconds to build instance. [ 944.732706] env[61868]: DEBUG oslo_concurrency.lockutils [req-6ba70ff4-8394-4d7c-9f31-3bad47a6af19 req-e17a05df-dd20-4e7b-84e2-6d6160fcf461 service nova] Releasing lock "refresh_cache-cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.732982] env[61868]: DEBUG nova.compute.manager [req-6ba70ff4-8394-4d7c-9f31-3bad47a6af19 req-e17a05df-dd20-4e7b-84e2-6d6160fcf461 service nova] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Received event network-vif-deleted-61f04f9e-9deb-496a-9301-5db736ced099 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 944.733272] env[61868]: INFO nova.compute.manager [req-6ba70ff4-8394-4d7c-9f31-3bad47a6af19 req-e17a05df-dd20-4e7b-84e2-6d6160fcf461 service nova] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Neutron deleted interface 61f04f9e-9deb-496a-9301-5db736ced099; detaching it from the instance and deleting it from the info cache [ 944.733631] env[61868]: DEBUG nova.network.neutron [req-6ba70ff4-8394-4d7c-9f31-3bad47a6af19 req-e17a05df-dd20-4e7b-84e2-6d6160fcf461 service nova] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.734734] env[61868]: DEBUG oslo_vmware.api [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315747, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.530988} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.736443] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 867384c6-175b-4848-b9ac-917470742ba2/867384c6-175b-4848-b9ac-917470742ba2.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 944.736773] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 944.737537] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b12fd8a-6371-47b7-a55c-e41e4ebcb9d4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.740394] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b0229c97-10ae-42db-bd13-1b6b1a2c0ddc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.750090] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dc6ef00-fefe-49ae-aaa9-6ba41723a772 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.754832] env[61868]: DEBUG oslo_vmware.api [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 944.754832] env[61868]: value = "task-1315749" [ 944.754832] env[61868]: _type = "Task" [ 944.754832] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.788162] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8dae21d-1b94-44ed-b0c4-17261d425ecc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.793634] env[61868]: DEBUG oslo_vmware.api [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315749, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.799514] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-321d3c71-5f68-4716-b23d-ddb3a9d6ddbc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.815355] env[61868]: DEBUG nova.compute.provider_tree [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 944.825836] env[61868]: DEBUG oslo_vmware.api [None req-d6dff523-f294-488f-8744-0eca0719bc34 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315738, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.881972] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315748, 'name': CreateVM_Task, 'duration_secs': 0.516669} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.882268] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 944.883254] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.883509] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.884030] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 944.884302] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39ae18c1-84e8-45c0-a6fb-01798ef00fb3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.891088] env[61868]: DEBUG oslo_vmware.api [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 944.891088] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52886327-87ea-ce57-7e39-5c7161d0648d" [ 944.891088] env[61868]: _type = "Task" [ 944.891088] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.903727] env[61868]: DEBUG oslo_vmware.api [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52886327-87ea-ce57-7e39-5c7161d0648d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.996764] env[61868]: DEBUG oslo_concurrency.lockutils [None req-294190a7-023a-4872-9a21-68724531565d tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.189903] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c567cdcb-0e3e-491c-9e6a-0961845dc92d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.192833] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d9bf584f-d218-42ed-b3a3-2ab5dae2b191 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Lock "87b25494-1853-4c7b-ae8b-6283992e895c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.839s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.210021] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c6e8a23-afe9-4417-aea3-9f4c87ef68b8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.217444] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Updating instance '842c8606-632a-47d4-9a67-38d8e6ce6b4d' progress to 83 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 945.236030] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b8d44342-3cbb-4636-914a-264d5e4ea9ae {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.246255] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eca5e494-1c07-465c-a095-a4f1d6d3605e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.265614] env[61868]: DEBUG oslo_vmware.api [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315749, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065081} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.265891] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 945.266655] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0ad5787-d5a6-42b3-8a74-9d7e553d4967 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.299531] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] 867384c6-175b-4848-b9ac-917470742ba2/867384c6-175b-4848-b9ac-917470742ba2.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 945.307911] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3ef683e9-a7e5-49db-8350-b21f82835f60 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.321812] env[61868]: DEBUG nova.compute.manager [req-6ba70ff4-8394-4d7c-9f31-3bad47a6af19 req-e17a05df-dd20-4e7b-84e2-6d6160fcf461 service nova] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Detach interface failed, port_id=61f04f9e-9deb-496a-9301-5db736ced099, reason: Instance 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9 could not be found. {{(pid=61868) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 945.323018] env[61868]: DEBUG nova.scheduler.client.report [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 945.342540] env[61868]: DEBUG oslo_vmware.api [None req-d6dff523-f294-488f-8744-0eca0719bc34 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315738, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.342847] env[61868]: DEBUG oslo_vmware.api [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 945.342847] env[61868]: value = "task-1315750" [ 945.342847] env[61868]: _type = "Task" [ 945.342847] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.355368] env[61868]: DEBUG oslo_vmware.api [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315750, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.403121] env[61868]: DEBUG oslo_vmware.api [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52886327-87ea-ce57-7e39-5c7161d0648d, 'name': SearchDatastore_Task, 'duration_secs': 0.051233} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.403503] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.403745] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 945.404024] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.404142] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.404330] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 945.404615] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f01aa3c0-a585-447b-b400-b561df925b1d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.414540] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 945.414731] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 945.415484] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-86412160-a324-4aa9-b815-c85ff6727318 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.423420] env[61868]: DEBUG oslo_vmware.api [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 945.423420] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]527eee86-56c3-b26a-0fe0-7e035cfe3cf4" [ 945.423420] env[61868]: _type = "Task" [ 945.423420] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.431858] env[61868]: DEBUG oslo_vmware.api [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]527eee86-56c3-b26a-0fe0-7e035cfe3cf4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.448910] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "4216e9b6-7b86-409e-9543-1c57e0f4b850" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.449189] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "4216e9b6-7b86-409e-9543-1c57e0f4b850" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.723705] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 945.724066] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ff9b9e26-cb40-4e58-9f16-1c7c0b5069fc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.732978] env[61868]: DEBUG oslo_vmware.api [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 945.732978] env[61868]: value = "task-1315751" [ 945.732978] env[61868]: _type = "Task" [ 945.732978] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.741915] env[61868]: DEBUG oslo_vmware.api [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315751, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.835722] env[61868]: DEBUG oslo_concurrency.lockutils [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.798s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.838309] env[61868]: DEBUG oslo_vmware.api [None req-d6dff523-f294-488f-8744-0eca0719bc34 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315738, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.839054] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d54e0102-d52b-44c3-93dd-ba46528a03fc tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.423s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.839054] env[61868]: DEBUG nova.objects.instance [None req-d54e0102-d52b-44c3-93dd-ba46528a03fc tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Lazy-loading 'resources' on Instance uuid 19c2720c-90bc-47f6-999b-6031f893408d {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 945.856408] env[61868]: DEBUG oslo_vmware.api [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315750, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.864676] env[61868]: INFO nova.scheduler.client.report [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Deleted allocations for instance b51e968e-6911-47f3-b5a9-e8a5865e42a0 [ 945.936485] env[61868]: DEBUG oslo_vmware.api [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]527eee86-56c3-b26a-0fe0-7e035cfe3cf4, 'name': SearchDatastore_Task, 'duration_secs': 0.009443} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.937263] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa8717e7-a81b-4fbd-b962-d30810fa965b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.944208] env[61868]: DEBUG oslo_vmware.api [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 945.944208] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52aa0f6f-40a1-d9be-75b7-5f847f89bf92" [ 945.944208] env[61868]: _type = "Task" [ 945.944208] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.953213] env[61868]: DEBUG nova.compute.manager [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 945.955851] env[61868]: DEBUG oslo_vmware.api [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52aa0f6f-40a1-d9be-75b7-5f847f89bf92, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.244370] env[61868]: DEBUG oslo_vmware.api [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315751, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.278506] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b062613f-01e0-48de-a1c7-661f529c1378 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Acquiring lock "87b25494-1853-4c7b-ae8b-6283992e895c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.278800] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b062613f-01e0-48de-a1c7-661f529c1378 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Lock "87b25494-1853-4c7b-ae8b-6283992e895c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.279028] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b062613f-01e0-48de-a1c7-661f529c1378 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Acquiring lock "87b25494-1853-4c7b-ae8b-6283992e895c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.279227] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b062613f-01e0-48de-a1c7-661f529c1378 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Lock "87b25494-1853-4c7b-ae8b-6283992e895c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.279402] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b062613f-01e0-48de-a1c7-661f529c1378 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Lock "87b25494-1853-4c7b-ae8b-6283992e895c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.281561] env[61868]: INFO nova.compute.manager [None req-b062613f-01e0-48de-a1c7-661f529c1378 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Terminating instance [ 946.283544] env[61868]: DEBUG nova.compute.manager [None req-b062613f-01e0-48de-a1c7-661f529c1378 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 946.283748] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-b062613f-01e0-48de-a1c7-661f529c1378 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 946.284596] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8fcedf8-448e-4a41-8d37-f118e377b884 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.292560] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-b062613f-01e0-48de-a1c7-661f529c1378 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 946.292802] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c3bf3beb-1f93-483f-b8fb-8a15e0e33c27 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.300486] env[61868]: DEBUG oslo_vmware.api [None req-b062613f-01e0-48de-a1c7-661f529c1378 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Waiting for the task: (returnval){ [ 946.300486] env[61868]: value = "task-1315753" [ 946.300486] env[61868]: _type = "Task" [ 946.300486] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.310859] env[61868]: DEBUG oslo_vmware.api [None req-b062613f-01e0-48de-a1c7-661f529c1378 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Task: {'id': task-1315753, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.334568] env[61868]: DEBUG oslo_vmware.api [None req-d6dff523-f294-488f-8744-0eca0719bc34 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315738, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.357623] env[61868]: DEBUG oslo_vmware.api [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315750, 'name': ReconfigVM_Task, 'duration_secs': 0.611539} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.357851] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Reconfigured VM instance instance-00000055 to attach disk [datastore1] 867384c6-175b-4848-b9ac-917470742ba2/867384c6-175b-4848-b9ac-917470742ba2.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 946.358472] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2afc5dd4-0174-4884-806e-7c4f338e463d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.367907] env[61868]: DEBUG oslo_vmware.api [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 946.367907] env[61868]: value = "task-1315754" [ 946.367907] env[61868]: _type = "Task" [ 946.367907] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.377727] env[61868]: DEBUG oslo_concurrency.lockutils [None req-deac1880-311b-43e1-a3f5-597780a704be tempest-ServerShowV254Test-1979036428 tempest-ServerShowV254Test-1979036428-project-member] Lock "b51e968e-6911-47f3-b5a9-e8a5865e42a0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.073s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.387532] env[61868]: DEBUG oslo_vmware.api [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315754, 'name': Rename_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.466687] env[61868]: DEBUG oslo_vmware.api [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52aa0f6f-40a1-d9be-75b7-5f847f89bf92, 'name': SearchDatastore_Task, 'duration_secs': 0.010892} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.469384] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.470160] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b/cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 946.470947] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-abede030-72ab-4827-9cf8-7ff47d3b7c18 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.485568] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.487330] env[61868]: DEBUG oslo_vmware.api [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 946.487330] env[61868]: value = "task-1315755" [ 946.487330] env[61868]: _type = "Task" [ 946.487330] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.500487] env[61868]: DEBUG oslo_vmware.api [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315755, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.579226] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e76cced6-1312-4766-b31a-93c7fc3a863a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.589349] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cdba347-6717-4234-b83b-9b33593f13eb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.621670] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab90d040-3a67-48b6-8acf-3d97429e2df1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.630358] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b1cc971-a1eb-4ce1-af9c-0d9361533b88 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.645686] env[61868]: DEBUG nova.compute.provider_tree [None req-d54e0102-d52b-44c3-93dd-ba46528a03fc tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 946.746429] env[61868]: DEBUG oslo_vmware.api [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315751, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.813786] env[61868]: DEBUG oslo_vmware.api [None req-b062613f-01e0-48de-a1c7-661f529c1378 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Task: {'id': task-1315753, 'name': PowerOffVM_Task, 'duration_secs': 0.222886} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.814012] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-b062613f-01e0-48de-a1c7-661f529c1378 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 946.814254] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-b062613f-01e0-48de-a1c7-661f529c1378 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 946.814544] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-63baf16c-0473-4964-9097-cd2b083e5e5b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.836034] env[61868]: DEBUG oslo_vmware.api [None req-d6dff523-f294-488f-8744-0eca0719bc34 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315738, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.883280] env[61868]: DEBUG oslo_vmware.api [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315754, 'name': Rename_Task, 'duration_secs': 0.188212} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.883582] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 946.883846] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9bcfe2fd-0dc9-4c5a-9c56-061941acb19e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.893878] env[61868]: DEBUG oslo_vmware.api [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 946.893878] env[61868]: value = "task-1315757" [ 946.893878] env[61868]: _type = "Task" [ 946.893878] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.900645] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-b062613f-01e0-48de-a1c7-661f529c1378 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 946.900878] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-b062613f-01e0-48de-a1c7-661f529c1378 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 946.901081] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-b062613f-01e0-48de-a1c7-661f529c1378 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Deleting the datastore file [datastore1] 87b25494-1853-4c7b-ae8b-6283992e895c {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 946.905647] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-560a2a31-089f-456d-9c6e-00005f6f538b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.907494] env[61868]: DEBUG oslo_vmware.api [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315757, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.914398] env[61868]: DEBUG oslo_vmware.api [None req-b062613f-01e0-48de-a1c7-661f529c1378 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Waiting for the task: (returnval){ [ 946.914398] env[61868]: value = "task-1315758" [ 946.914398] env[61868]: _type = "Task" [ 946.914398] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.927806] env[61868]: DEBUG oslo_vmware.api [None req-b062613f-01e0-48de-a1c7-661f529c1378 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Task: {'id': task-1315758, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.998374] env[61868]: DEBUG oslo_vmware.api [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315755, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.151775] env[61868]: DEBUG nova.scheduler.client.report [None req-d54e0102-d52b-44c3-93dd-ba46528a03fc tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 947.246347] env[61868]: DEBUG oslo_vmware.api [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315751, 'name': PowerOnVM_Task, 'duration_secs': 1.141444} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.246635] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 947.246852] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1af9d10d-59c6-48a8-b399-e16e3823bf56 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Updating instance '842c8606-632a-47d4-9a67-38d8e6ce6b4d' progress to 100 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 947.339447] env[61868]: DEBUG oslo_vmware.api [None req-d6dff523-f294-488f-8744-0eca0719bc34 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315738, 'name': ReconfigVM_Task, 'duration_secs': 5.789506} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.339825] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d6dff523-f294-488f-8744-0eca0719bc34 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Releasing lock "5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.340120] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-d6dff523-f294-488f-8744-0eca0719bc34 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Reconfigured VM to detach interface {{(pid=61868) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 947.405233] env[61868]: DEBUG oslo_vmware.api [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315757, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.424936] env[61868]: DEBUG oslo_vmware.api [None req-b062613f-01e0-48de-a1c7-661f529c1378 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Task: {'id': task-1315758, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.215396} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.425229] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-b062613f-01e0-48de-a1c7-661f529c1378 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 947.425422] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-b062613f-01e0-48de-a1c7-661f529c1378 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 947.425666] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-b062613f-01e0-48de-a1c7-661f529c1378 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 947.425956] env[61868]: INFO nova.compute.manager [None req-b062613f-01e0-48de-a1c7-661f529c1378 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Took 1.14 seconds to destroy the instance on the hypervisor. [ 947.426297] env[61868]: DEBUG oslo.service.loopingcall [None req-b062613f-01e0-48de-a1c7-661f529c1378 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 947.426520] env[61868]: DEBUG nova.compute.manager [-] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 947.426617] env[61868]: DEBUG nova.network.neutron [-] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 947.499468] env[61868]: DEBUG oslo_vmware.api [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315755, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.516857} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.499926] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b/cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 947.500324] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 947.500703] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2db8354f-e407-4095-9be0-2971ffde5caa {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.511019] env[61868]: DEBUG oslo_vmware.api [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 947.511019] env[61868]: value = "task-1315759" [ 947.511019] env[61868]: _type = "Task" [ 947.511019] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.518026] env[61868]: DEBUG oslo_vmware.api [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315759, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.661508] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d54e0102-d52b-44c3-93dd-ba46528a03fc tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.823s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.664234] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0b461ce5-034e-41b9-8c34-be9526d11ff6 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.323s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.664535] env[61868]: DEBUG nova.objects.instance [None req-0b461ce5-034e-41b9-8c34-be9526d11ff6 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Lazy-loading 'resources' on Instance uuid 6ce39c07-939c-49f3-8871-6e2163709f4a {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 947.692868] env[61868]: INFO nova.scheduler.client.report [None req-d54e0102-d52b-44c3-93dd-ba46528a03fc tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Deleted allocations for instance 19c2720c-90bc-47f6-999b-6031f893408d [ 947.760423] env[61868]: DEBUG nova.compute.manager [req-fc80b321-b533-4a0d-be3c-196fcfbe2535 req-9cb378a4-02aa-4e53-8bdd-d8e72c9b9739 service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Received event network-vif-deleted-8a5f760c-c14e-4701-bde0-97fff3b6289e {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 947.760633] env[61868]: INFO nova.compute.manager [req-fc80b321-b533-4a0d-be3c-196fcfbe2535 req-9cb378a4-02aa-4e53-8bdd-d8e72c9b9739 service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Neutron deleted interface 8a5f760c-c14e-4701-bde0-97fff3b6289e; detaching it from the instance and deleting it from the info cache [ 947.760917] env[61868]: DEBUG nova.network.neutron [req-fc80b321-b533-4a0d-be3c-196fcfbe2535 req-9cb378a4-02aa-4e53-8bdd-d8e72c9b9739 service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Updating instance_info_cache with network_info: [{"id": "9771cedf-dae3-41e8-ad4b-aec421274d6c", "address": "fa:16:3e:4f:5f:b8", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.244", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9771cedf-da", "ovs_interfaceid": "9771cedf-dae3-41e8-ad4b-aec421274d6c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "87455df2-50af-4a63-b58b-4813a20d0b38", "address": "fa:16:3e:51:bc:03", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87455df2-50", "ovs_interfaceid": "87455df2-50af-4a63-b58b-4813a20d0b38", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.904830] env[61868]: DEBUG oslo_vmware.api [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315757, 'name': PowerOnVM_Task, 'duration_secs': 0.519962} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.905235] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 947.905357] env[61868]: INFO nova.compute.manager [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Took 8.79 seconds to spawn the instance on the hypervisor. [ 947.905562] env[61868]: DEBUG nova.compute.manager [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 947.906362] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec55df19-a3e4-417e-a244-d3efea5a7ef8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.021347] env[61868]: DEBUG oslo_vmware.api [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315759, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069762} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.021347] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 948.021580] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54e9e9c1-bd25-44f0-9c09-6123e0a7c988 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.046214] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b/cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 948.046532] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a5d5af90-a65e-4e6d-b346-fbf3c4b04e28 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.069102] env[61868]: DEBUG oslo_vmware.api [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 948.069102] env[61868]: value = "task-1315760" [ 948.069102] env[61868]: _type = "Task" [ 948.069102] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.077921] env[61868]: DEBUG oslo_vmware.api [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315760, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.203964] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d54e0102-d52b-44c3-93dd-ba46528a03fc tempest-ServerRescueTestJSON-1862169840 tempest-ServerRescueTestJSON-1862169840-project-member] Lock "19c2720c-90bc-47f6-999b-6031f893408d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.151s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.264437] env[61868]: DEBUG oslo_concurrency.lockutils [req-fc80b321-b533-4a0d-be3c-196fcfbe2535 req-9cb378a4-02aa-4e53-8bdd-d8e72c9b9739 service nova] Acquiring lock "5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.264636] env[61868]: DEBUG oslo_concurrency.lockutils [req-fc80b321-b533-4a0d-be3c-196fcfbe2535 req-9cb378a4-02aa-4e53-8bdd-d8e72c9b9739 service nova] Acquired lock "5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.266975] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e914b0e4-ebeb-4b41-a6a3-232e1f39f6ec {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.293823] env[61868]: DEBUG nova.network.neutron [-] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.295100] env[61868]: DEBUG oslo_concurrency.lockutils [req-fc80b321-b533-4a0d-be3c-196fcfbe2535 req-9cb378a4-02aa-4e53-8bdd-d8e72c9b9739 service nova] Releasing lock "5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.295397] env[61868]: WARNING nova.compute.manager [req-fc80b321-b533-4a0d-be3c-196fcfbe2535 req-9cb378a4-02aa-4e53-8bdd-d8e72c9b9739 service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Detach interface failed, port_id=8a5f760c-c14e-4701-bde0-97fff3b6289e, reason: No device with interface-id 8a5f760c-c14e-4701-bde0-97fff3b6289e exists on VM: nova.exception.NotFound: No device with interface-id 8a5f760c-c14e-4701-bde0-97fff3b6289e exists on VM [ 948.342765] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efd3918f-44ef-417a-81bb-3e2f603df5e4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.353147] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7db8359c-3356-47ae-9a3d-6ad893cc64a7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.394119] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75a58483-6df7-4a5d-8cec-65eb86ff2c9f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.409624] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-388fd304-2641-46e4-b87d-af99c800a4a2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.427216] env[61868]: INFO nova.compute.manager [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Took 29.86 seconds to build instance. [ 948.437036] env[61868]: DEBUG nova.compute.provider_tree [None req-0b461ce5-034e-41b9-8c34-be9526d11ff6 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 948.584623] env[61868]: DEBUG oslo_vmware.api [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315760, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.730347] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d6dff523-f294-488f-8744-0eca0719bc34 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "refresh_cache-5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.730670] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d6dff523-f294-488f-8744-0eca0719bc34 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquired lock "refresh_cache-5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.730987] env[61868]: DEBUG nova.network.neutron [None req-d6dff523-f294-488f-8744-0eca0719bc34 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 948.796493] env[61868]: INFO nova.compute.manager [-] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Took 1.37 seconds to deallocate network for instance. [ 948.886842] env[61868]: DEBUG nova.compute.manager [req-0f075bba-e248-47ea-a127-4f37ec996e15 req-8892b0b6-2f80-45f9-8c0d-d5c112c3dbe8 service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Received event network-vif-deleted-87455df2-50af-4a63-b58b-4813a20d0b38 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 948.887077] env[61868]: INFO nova.compute.manager [req-0f075bba-e248-47ea-a127-4f37ec996e15 req-8892b0b6-2f80-45f9-8c0d-d5c112c3dbe8 service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Neutron deleted interface 87455df2-50af-4a63-b58b-4813a20d0b38; detaching it from the instance and deleting it from the info cache [ 948.887334] env[61868]: DEBUG nova.network.neutron [req-0f075bba-e248-47ea-a127-4f37ec996e15 req-8892b0b6-2f80-45f9-8c0d-d5c112c3dbe8 service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Updating instance_info_cache with network_info: [{"id": "9771cedf-dae3-41e8-ad4b-aec421274d6c", "address": "fa:16:3e:4f:5f:b8", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.244", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9771cedf-da", "ovs_interfaceid": "9771cedf-dae3-41e8-ad4b-aec421274d6c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.903880] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48267cd4-153f-4eea-b415-668e133fc799 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.913341] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-2e090915-872c-4850-93f5-3c73aa893534 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Suspending the VM {{(pid=61868) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 948.914807] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-08c963ce-d807-4b89-b1aa-d7ddce1b9721 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.921871] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquiring lock "75fdbc11-1ef3-442d-bcf4-55069ba9ab62" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.922156] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lock "75fdbc11-1ef3-442d-bcf4-55069ba9ab62" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.922347] env[61868]: INFO nova.compute.manager [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Shelving [ 948.923827] env[61868]: DEBUG oslo_vmware.api [None req-2e090915-872c-4850-93f5-3c73aa893534 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 948.923827] env[61868]: value = "task-1315761" [ 948.923827] env[61868]: _type = "Task" [ 948.923827] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.935616] env[61868]: DEBUG oslo_vmware.api [None req-2e090915-872c-4850-93f5-3c73aa893534 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315761, 'name': SuspendVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.939136] env[61868]: DEBUG oslo_concurrency.lockutils [None req-42008d2d-b718-4585-b02f-14670938e780 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "867384c6-175b-4848-b9ac-917470742ba2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.380s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.940173] env[61868]: DEBUG nova.scheduler.client.report [None req-0b461ce5-034e-41b9-8c34-be9526d11ff6 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 949.085517] env[61868]: DEBUG oslo_vmware.api [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315760, 'name': ReconfigVM_Task, 'duration_secs': 0.648817} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.085918] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Reconfigured VM instance instance-00000056 to attach disk [datastore1] cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b/cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 949.086529] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e02f4c9f-b60a-41b5-ab95-b65a42020125 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.097712] env[61868]: DEBUG oslo_vmware.api [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 949.097712] env[61868]: value = "task-1315762" [ 949.097712] env[61868]: _type = "Task" [ 949.097712] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.115348] env[61868]: DEBUG oslo_vmware.api [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315762, 'name': Rename_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.307350] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b062613f-01e0-48de-a1c7-661f529c1378 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.362547] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f65c85db-097a-4399-ba19-932a22223218 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "5aec2609-79d3-4725-a182-70b575adbe17" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.362547] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f65c85db-097a-4399-ba19-932a22223218 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "5aec2609-79d3-4725-a182-70b575adbe17" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.362547] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f65c85db-097a-4399-ba19-932a22223218 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "5aec2609-79d3-4725-a182-70b575adbe17-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.362547] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f65c85db-097a-4399-ba19-932a22223218 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "5aec2609-79d3-4725-a182-70b575adbe17-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.362547] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f65c85db-097a-4399-ba19-932a22223218 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "5aec2609-79d3-4725-a182-70b575adbe17-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.366937] env[61868]: INFO nova.compute.manager [None req-f65c85db-097a-4399-ba19-932a22223218 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Terminating instance [ 949.369052] env[61868]: DEBUG nova.compute.manager [None req-f65c85db-097a-4399-ba19-932a22223218 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 949.369686] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f65c85db-097a-4399-ba19-932a22223218 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 949.370073] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1625002f-0b88-416f-8360-f808d6cd2562 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.384831] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-f65c85db-097a-4399-ba19-932a22223218 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 949.385145] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-66cf31f3-d920-4adb-a185-e4b9774ec8c5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.392258] env[61868]: DEBUG oslo_concurrency.lockutils [req-0f075bba-e248-47ea-a127-4f37ec996e15 req-8892b0b6-2f80-45f9-8c0d-d5c112c3dbe8 service nova] Acquiring lock "5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.393675] env[61868]: DEBUG oslo_vmware.api [None req-f65c85db-097a-4399-ba19-932a22223218 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for the task: (returnval){ [ 949.393675] env[61868]: value = "task-1315763" [ 949.393675] env[61868]: _type = "Task" [ 949.393675] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.406282] env[61868]: DEBUG oslo_vmware.api [None req-f65c85db-097a-4399-ba19-932a22223218 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315763, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.436015] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 949.436326] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e0f798ad-d5e2-4b48-9cdb-16b6cdba481f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.442418] env[61868]: DEBUG oslo_vmware.api [None req-2e090915-872c-4850-93f5-3c73aa893534 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315761, 'name': SuspendVM_Task} progress is 58%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.443934] env[61868]: DEBUG oslo_vmware.api [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the task: (returnval){ [ 949.443934] env[61868]: value = "task-1315764" [ 949.443934] env[61868]: _type = "Task" [ 949.443934] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.444828] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0b461ce5-034e-41b9-8c34-be9526d11ff6 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.781s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.447593] env[61868]: DEBUG oslo_concurrency.lockutils [None req-294190a7-023a-4872-9a21-68724531565d tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.451s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.447866] env[61868]: DEBUG nova.objects.instance [None req-294190a7-023a-4872-9a21-68724531565d tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Lazy-loading 'resources' on Instance uuid 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 949.460838] env[61868]: DEBUG oslo_vmware.api [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315764, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.474042] env[61868]: INFO nova.scheduler.client.report [None req-0b461ce5-034e-41b9-8c34-be9526d11ff6 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Deleted allocations for instance 6ce39c07-939c-49f3-8871-6e2163709f4a [ 949.610162] env[61868]: DEBUG oslo_vmware.api [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315762, 'name': Rename_Task, 'duration_secs': 0.260397} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.610556] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 949.611281] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0b16a361-eca7-429f-a37b-d5a13bea99cd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.615675] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e22cc326-cf74-4114-bbe4-7c54f0664481 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "842c8606-632a-47d4-9a67-38d8e6ce6b4d" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.615973] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e22cc326-cf74-4114-bbe4-7c54f0664481 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "842c8606-632a-47d4-9a67-38d8e6ce6b4d" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.616257] env[61868]: DEBUG nova.compute.manager [None req-e22cc326-cf74-4114-bbe4-7c54f0664481 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Going to confirm migration 1 {{(pid=61868) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 949.625026] env[61868]: DEBUG oslo_vmware.api [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 949.625026] env[61868]: value = "task-1315765" [ 949.625026] env[61868]: _type = "Task" [ 949.625026] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.635992] env[61868]: DEBUG oslo_vmware.api [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315765, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.850976] env[61868]: DEBUG nova.compute.manager [req-d90e3401-47e7-4de6-acfd-2fafd7b7aaaf req-b9e6875e-9d18-4eab-bb89-87d1b0a50ef4 service nova] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Received event network-vif-deleted-bbcb967e-bf16-4589-ab6a-23e5ce9e8dd4 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 949.906363] env[61868]: DEBUG oslo_vmware.api [None req-f65c85db-097a-4399-ba19-932a22223218 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315763, 'name': PowerOffVM_Task, 'duration_secs': 0.251364} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.906685] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-f65c85db-097a-4399-ba19-932a22223218 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 949.906864] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f65c85db-097a-4399-ba19-932a22223218 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 949.907096] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-20875b51-2ec2-4737-b384-44221d30ae1f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.909903] env[61868]: INFO nova.network.neutron [None req-d6dff523-f294-488f-8744-0eca0719bc34 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Port 87455df2-50af-4a63-b58b-4813a20d0b38 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 949.910249] env[61868]: DEBUG nova.network.neutron [None req-d6dff523-f294-488f-8744-0eca0719bc34 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Updating instance_info_cache with network_info: [{"id": "9771cedf-dae3-41e8-ad4b-aec421274d6c", "address": "fa:16:3e:4f:5f:b8", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.244", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9771cedf-da", "ovs_interfaceid": "9771cedf-dae3-41e8-ad4b-aec421274d6c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.945907] env[61868]: DEBUG oslo_vmware.api [None req-2e090915-872c-4850-93f5-3c73aa893534 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315761, 'name': SuspendVM_Task, 'duration_secs': 0.656031} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.946667] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-2e090915-872c-4850-93f5-3c73aa893534 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Suspended the VM {{(pid=61868) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 949.947172] env[61868]: DEBUG nova.compute.manager [None req-2e090915-872c-4850-93f5-3c73aa893534 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 949.948322] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b9c3d49-e39a-40bc-a4b1-64f315affb5f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.978656] env[61868]: DEBUG oslo_vmware.api [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315764, 'name': PowerOffVM_Task, 'duration_secs': 0.238695} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.979226] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 949.980079] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be517f99-807c-4fe3-b266-f7b8df052167 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.990173] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0b461ce5-034e-41b9-8c34-be9526d11ff6 tempest-ServerAddressesNegativeTestJSON-977709259 tempest-ServerAddressesNegativeTestJSON-977709259-project-member] Lock "6ce39c07-939c-49f3-8871-6e2163709f4a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.632s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.011375] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7643e295-2861-4f7e-be6d-68c1a93b5719 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.061956] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f65c85db-097a-4399-ba19-932a22223218 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 950.061956] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f65c85db-097a-4399-ba19-932a22223218 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Deleting contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 950.061956] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-f65c85db-097a-4399-ba19-932a22223218 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Deleting the datastore file [datastore2] 5aec2609-79d3-4725-a182-70b575adbe17 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 950.061956] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e4b2de91-5ddf-4f05-af68-5a0d3c2f0bd8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.074026] env[61868]: DEBUG oslo_vmware.api [None req-f65c85db-097a-4399-ba19-932a22223218 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for the task: (returnval){ [ 950.074026] env[61868]: value = "task-1315767" [ 950.074026] env[61868]: _type = "Task" [ 950.074026] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.085919] env[61868]: DEBUG oslo_vmware.api [None req-f65c85db-097a-4399-ba19-932a22223218 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315767, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.136332] env[61868]: DEBUG oslo_vmware.api [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315765, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.156140] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae505cf0-5f2d-45c2-8f0c-43a57209ad02 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.164239] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ee14571-69af-4892-b9e8-132752fdbdee {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.197541] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccf5224b-3bf7-4449-9d28-92cb8f7348c3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.205745] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f929a37-32f0-4b36-ad4f-73d14a9a45c5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.220364] env[61868]: DEBUG nova.compute.provider_tree [None req-294190a7-023a-4872-9a21-68724531565d tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 950.248811] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e22cc326-cf74-4114-bbe4-7c54f0664481 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "refresh_cache-842c8606-632a-47d4-9a67-38d8e6ce6b4d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.249737] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e22cc326-cf74-4114-bbe4-7c54f0664481 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquired lock "refresh_cache-842c8606-632a-47d4-9a67-38d8e6ce6b4d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.249737] env[61868]: DEBUG nova.network.neutron [None req-e22cc326-cf74-4114-bbe4-7c54f0664481 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 950.249737] env[61868]: DEBUG nova.objects.instance [None req-e22cc326-cf74-4114-bbe4-7c54f0664481 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lazy-loading 'info_cache' on Instance uuid 842c8606-632a-47d4-9a67-38d8e6ce6b4d {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 950.415177] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d6dff523-f294-488f-8744-0eca0719bc34 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Releasing lock "refresh_cache-5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.525774] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Creating Snapshot of the VM instance {{(pid=61868) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 950.527491] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-631bcf78-d906-44c9-ba04-af34042e4fb2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.539090] env[61868]: DEBUG oslo_vmware.api [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the task: (returnval){ [ 950.539090] env[61868]: value = "task-1315768" [ 950.539090] env[61868]: _type = "Task" [ 950.539090] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.554837] env[61868]: DEBUG oslo_vmware.api [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315768, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.588518] env[61868]: DEBUG oslo_vmware.api [None req-f65c85db-097a-4399-ba19-932a22223218 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315767, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.164923} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.588518] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-f65c85db-097a-4399-ba19-932a22223218 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 950.588518] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f65c85db-097a-4399-ba19-932a22223218 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Deleted contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 950.588518] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f65c85db-097a-4399-ba19-932a22223218 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 950.588518] env[61868]: INFO nova.compute.manager [None req-f65c85db-097a-4399-ba19-932a22223218 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Took 1.22 seconds to destroy the instance on the hypervisor. [ 950.588518] env[61868]: DEBUG oslo.service.loopingcall [None req-f65c85db-097a-4399-ba19-932a22223218 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 950.588518] env[61868]: DEBUG nova.compute.manager [-] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 950.588792] env[61868]: DEBUG nova.network.neutron [-] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 950.637948] env[61868]: DEBUG oslo_vmware.api [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315765, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.698759] env[61868]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port 87455df2-50af-4a63-b58b-4813a20d0b38 could not be found.", "detail": ""}} {{(pid=61868) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 950.699312] env[61868]: DEBUG nova.network.neutron [-] Unable to show port 87455df2-50af-4a63-b58b-4813a20d0b38 as it no longer exists. {{(pid=61868) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 950.723403] env[61868]: DEBUG nova.scheduler.client.report [None req-294190a7-023a-4872-9a21-68724531565d tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 950.924523] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d6dff523-f294-488f-8744-0eca0719bc34 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "interface-5aec2609-79d3-4725-a182-70b575adbe17-8a5f760c-c14e-4701-bde0-97fff3b6289e" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.214s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.051712] env[61868]: DEBUG oslo_vmware.api [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315768, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.138233] env[61868]: DEBUG oslo_vmware.api [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315765, 'name': PowerOnVM_Task, 'duration_secs': 1.082308} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.138233] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 951.138233] env[61868]: INFO nova.compute.manager [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Took 9.70 seconds to spawn the instance on the hypervisor. [ 951.138387] env[61868]: DEBUG nova.compute.manager [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 951.139202] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a26297c-3168-4d28-a225-f1c5ac37c327 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.231813] env[61868]: DEBUG oslo_concurrency.lockutils [None req-294190a7-023a-4872-9a21-68724531565d tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.782s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.233864] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.748s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.235295] env[61868]: INFO nova.compute.claims [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 951.259813] env[61868]: INFO nova.scheduler.client.report [None req-294190a7-023a-4872-9a21-68724531565d tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Deleted allocations for instance 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9 [ 951.555550] env[61868]: DEBUG oslo_vmware.api [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315768, 'name': CreateSnapshot_Task, 'duration_secs': 0.90878} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.556537] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Created Snapshot of the VM instance {{(pid=61868) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 951.557454] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-820c4447-19ed-47a4-92a6-a890b3da9558 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.659025] env[61868]: DEBUG nova.network.neutron [None req-e22cc326-cf74-4114-bbe4-7c54f0664481 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Updating instance_info_cache with network_info: [{"id": "e3943fa9-423d-4168-8aa6-1651280cd127", "address": "fa:16:3e:1d:68:a3", "network": {"id": "c1f5bf03-438f-4acc-93fd-159bed7d4069", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2061742146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a60c3c20950e4156b64c3b4c61b9f0f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24b91e3b-50f1-4a16-b929-942e6b31b2d8", "external-id": "nsx-vlan-transportzone-34", "segmentation_id": 34, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3943fa9-42", "ovs_interfaceid": "e3943fa9-423d-4168-8aa6-1651280cd127", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.659647] env[61868]: INFO nova.compute.manager [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Took 24.58 seconds to build instance. [ 951.773185] env[61868]: DEBUG oslo_concurrency.lockutils [None req-294190a7-023a-4872-9a21-68724531565d tempest-ServerTagsTestJSON-2066012797 tempest-ServerTagsTestJSON-2066012797-project-member] Lock "4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.763s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.879709] env[61868]: DEBUG nova.network.neutron [-] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.891542] env[61868]: DEBUG nova.compute.manager [req-62f43b6a-035e-4c92-8209-28cacb21355c req-b19ebbca-f073-4cd1-a1af-1d4d4b5864e5 service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Received event network-vif-deleted-9771cedf-dae3-41e8-ad4b-aec421274d6c {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 951.893704] env[61868]: INFO nova.compute.manager [req-62f43b6a-035e-4c92-8209-28cacb21355c req-b19ebbca-f073-4cd1-a1af-1d4d4b5864e5 service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Neutron deleted interface 9771cedf-dae3-41e8-ad4b-aec421274d6c; detaching it from the instance and deleting it from the info cache [ 951.894594] env[61868]: DEBUG nova.network.neutron [req-62f43b6a-035e-4c92-8209-28cacb21355c req-b19ebbca-f073-4cd1-a1af-1d4d4b5864e5 service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.079994] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Creating linked-clone VM from snapshot {{(pid=61868) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 952.081091] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-2aca1086-7f5a-4761-a1c7-e2c54f2a0a83 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.091270] env[61868]: DEBUG oslo_vmware.api [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the task: (returnval){ [ 952.091270] env[61868]: value = "task-1315769" [ 952.091270] env[61868]: _type = "Task" [ 952.091270] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.102158] env[61868]: DEBUG oslo_vmware.api [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315769, 'name': CloneVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.162510] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e22cc326-cf74-4114-bbe4-7c54f0664481 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Releasing lock "refresh_cache-842c8606-632a-47d4-9a67-38d8e6ce6b4d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.162510] env[61868]: DEBUG nova.objects.instance [None req-e22cc326-cf74-4114-bbe4-7c54f0664481 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lazy-loading 'migration_context' on Instance uuid 842c8606-632a-47d4-9a67-38d8e6ce6b4d {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 952.163253] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8d2a1fc9-bd32-4ee0-8c8a-c798bcedcb6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.095s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.383235] env[61868]: INFO nova.compute.manager [-] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Took 1.79 seconds to deallocate network for instance. [ 952.397594] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-383bce7a-d681-47bf-86c3-7b9dc6b413e6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.410460] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1394a801-910b-47e1-9a5e-7f915ae99b41 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.457235] env[61868]: DEBUG nova.compute.manager [req-62f43b6a-035e-4c92-8209-28cacb21355c req-b19ebbca-f073-4cd1-a1af-1d4d4b5864e5 service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Detach interface failed, port_id=9771cedf-dae3-41e8-ad4b-aec421274d6c, reason: Instance 5aec2609-79d3-4725-a182-70b575adbe17 could not be found. {{(pid=61868) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 952.480471] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2231d30-2131-4252-af41-84a903b14033 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.490167] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fa01b8f-c204-40f6-9285-c67f624ffdeb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.531739] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a485d6a7-9f9a-45c3-a6b2-bb26e0874a0b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.539944] env[61868]: DEBUG nova.compute.manager [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 952.540795] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b46d1dcc-bcb3-4679-8d62-a58e9373f1bb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.544555] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3667933-5069-40d2-b7c5-b3cbe1c4a660 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.560141] env[61868]: DEBUG nova.compute.provider_tree [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 952.603816] env[61868]: DEBUG oslo_vmware.api [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315769, 'name': CloneVM_Task} progress is 94%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.664621] env[61868]: DEBUG nova.objects.base [None req-e22cc326-cf74-4114-bbe4-7c54f0664481 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Object Instance<842c8606-632a-47d4-9a67-38d8e6ce6b4d> lazy-loaded attributes: info_cache,migration_context {{(pid=61868) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 952.665947] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d958cf21-a6ee-467b-a657-3e07aaab0631 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.687581] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-059ab9e6-63e3-4184-b2e4-7463a032d1fa {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.696619] env[61868]: DEBUG oslo_vmware.api [None req-e22cc326-cf74-4114-bbe4-7c54f0664481 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 952.696619] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]521cd0b0-8f51-7e1f-81d1-58ddfb3da6c6" [ 952.696619] env[61868]: _type = "Task" [ 952.696619] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.706676] env[61868]: DEBUG oslo_vmware.api [None req-e22cc326-cf74-4114-bbe4-7c54f0664481 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]521cd0b0-8f51-7e1f-81d1-58ddfb3da6c6, 'name': SearchDatastore_Task, 'duration_secs': 0.007474} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.706965] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e22cc326-cf74-4114-bbe4-7c54f0664481 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.895229] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f65c85db-097a-4399-ba19-932a22223218 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.066539] env[61868]: DEBUG nova.scheduler.client.report [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 953.070781] env[61868]: INFO nova.compute.manager [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] instance snapshotting [ 953.071045] env[61868]: WARNING nova.compute.manager [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 953.073757] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9671ef8-597b-49de-91be-522b73d6fec0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.098381] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26f7eb84-1b87-441f-9b5f-dda6fe652aa7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.106990] env[61868]: DEBUG oslo_vmware.api [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315769, 'name': CloneVM_Task} progress is 94%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.250798] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d5307f8-13dc-420d-8f26-1559a2c96086 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.251586] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d5307f8-13dc-420d-8f26-1559a2c96086 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.251586] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d5307f8-13dc-420d-8f26-1559a2c96086 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.251586] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d5307f8-13dc-420d-8f26-1559a2c96086 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.251840] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d5307f8-13dc-420d-8f26-1559a2c96086 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.255025] env[61868]: INFO nova.compute.manager [None req-1d5307f8-13dc-420d-8f26-1559a2c96086 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Terminating instance [ 953.257077] env[61868]: DEBUG nova.compute.manager [None req-1d5307f8-13dc-420d-8f26-1559a2c96086 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 953.257346] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1d5307f8-13dc-420d-8f26-1559a2c96086 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 953.258373] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2939f04-5059-440d-ae91-ec4dd8577bd0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.281968] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d5307f8-13dc-420d-8f26-1559a2c96086 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 953.282362] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-724a49f8-e4b4-476c-9e0d-6f8824bdf975 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.291313] env[61868]: DEBUG oslo_vmware.api [None req-1d5307f8-13dc-420d-8f26-1559a2c96086 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 953.291313] env[61868]: value = "task-1315770" [ 953.291313] env[61868]: _type = "Task" [ 953.291313] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.301303] env[61868]: DEBUG oslo_vmware.api [None req-1d5307f8-13dc-420d-8f26-1559a2c96086 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315770, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.574629] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.341s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.575226] env[61868]: DEBUG nova.compute.manager [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 953.577943] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b062613f-01e0-48de-a1c7-661f529c1378 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.271s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.579099] env[61868]: DEBUG nova.objects.instance [None req-b062613f-01e0-48de-a1c7-661f529c1378 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Lazy-loading 'resources' on Instance uuid 87b25494-1853-4c7b-ae8b-6283992e895c {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 953.605419] env[61868]: DEBUG oslo_vmware.api [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315769, 'name': CloneVM_Task, 'duration_secs': 1.368422} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.605707] env[61868]: INFO nova.virt.vmwareapi.vmops [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Created linked-clone VM from snapshot [ 953.606527] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54e914f6-cf40-491c-b1f4-ed7beca61a0e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.615735] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Creating Snapshot of the VM instance {{(pid=61868) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 953.616020] env[61868]: DEBUG nova.virt.vmwareapi.images [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Uploading image 76af0b0a-47a5-487a-9b10-d16668b6caae {{(pid=61868) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 953.618565] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-4dfbb8d6-d40c-43ad-a669-be84b14660d2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.626985] env[61868]: DEBUG oslo_vmware.api [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 953.626985] env[61868]: value = "task-1315771" [ 953.626985] env[61868]: _type = "Task" [ 953.626985] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.636117] env[61868]: DEBUG oslo_vmware.api [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315771, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.644316] env[61868]: DEBUG oslo_vmware.rw_handles [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 953.644316] env[61868]: value = "vm-281591" [ 953.644316] env[61868]: _type = "VirtualMachine" [ 953.644316] env[61868]: }. {{(pid=61868) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 953.644587] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-31eeb875-bccc-41b8-986d-b317e7f2aeb5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.652493] env[61868]: DEBUG oslo_vmware.rw_handles [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lease: (returnval){ [ 953.652493] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e9b51d-8942-1302-5cfa-948fdec705e1" [ 953.652493] env[61868]: _type = "HttpNfcLease" [ 953.652493] env[61868]: } obtained for exporting VM: (result){ [ 953.652493] env[61868]: value = "vm-281591" [ 953.652493] env[61868]: _type = "VirtualMachine" [ 953.652493] env[61868]: }. {{(pid=61868) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 953.652840] env[61868]: DEBUG oslo_vmware.api [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the lease: (returnval){ [ 953.652840] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e9b51d-8942-1302-5cfa-948fdec705e1" [ 953.652840] env[61868]: _type = "HttpNfcLease" [ 953.652840] env[61868]: } to be ready. {{(pid=61868) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 953.660831] env[61868]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 953.660831] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e9b51d-8942-1302-5cfa-948fdec705e1" [ 953.660831] env[61868]: _type = "HttpNfcLease" [ 953.660831] env[61868]: } is initializing. {{(pid=61868) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 953.802822] env[61868]: DEBUG oslo_vmware.api [None req-1d5307f8-13dc-420d-8f26-1559a2c96086 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315770, 'name': PowerOffVM_Task, 'duration_secs': 0.251372} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.803274] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d5307f8-13dc-420d-8f26-1559a2c96086 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 953.803516] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1d5307f8-13dc-420d-8f26-1559a2c96086 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 953.803806] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d1c7dbfc-8cda-40f4-b429-ef58b927a012 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.870128] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1d5307f8-13dc-420d-8f26-1559a2c96086 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 953.870333] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1d5307f8-13dc-420d-8f26-1559a2c96086 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 953.870524] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d5307f8-13dc-420d-8f26-1559a2c96086 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Deleting the datastore file [datastore1] cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 953.870816] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-12bf74c8-62ab-49f3-a7a5-7d42adc9c79b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.878259] env[61868]: DEBUG oslo_vmware.api [None req-1d5307f8-13dc-420d-8f26-1559a2c96086 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 953.878259] env[61868]: value = "task-1315774" [ 953.878259] env[61868]: _type = "Task" [ 953.878259] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.886534] env[61868]: DEBUG oslo_vmware.api [None req-1d5307f8-13dc-420d-8f26-1559a2c96086 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315774, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.084482] env[61868]: DEBUG nova.compute.utils [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 954.088117] env[61868]: DEBUG nova.compute.manager [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 954.088348] env[61868]: DEBUG nova.network.neutron [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 954.142822] env[61868]: DEBUG oslo_vmware.api [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315771, 'name': CreateSnapshot_Task, 'duration_secs': 0.510723} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.143204] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Created Snapshot of the VM instance {{(pid=61868) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 954.144649] env[61868]: DEBUG nova.policy [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fd6ed09d33f6491b83dc72cef31bfd29', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '812e115cc1114d0e8536924c4f5556f8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 954.146679] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e9735da-3060-40b9-85bd-c0d5b41c8ce1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.168980] env[61868]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 954.168980] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e9b51d-8942-1302-5cfa-948fdec705e1" [ 954.168980] env[61868]: _type = "HttpNfcLease" [ 954.168980] env[61868]: } is ready. {{(pid=61868) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 954.171959] env[61868]: DEBUG oslo_vmware.rw_handles [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 954.171959] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e9b51d-8942-1302-5cfa-948fdec705e1" [ 954.171959] env[61868]: _type = "HttpNfcLease" [ 954.171959] env[61868]: }. {{(pid=61868) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 954.173208] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4d42d3b-b856-4a19-b70b-cd5ad8916332 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.185754] env[61868]: DEBUG oslo_vmware.rw_handles [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52bef67d-356c-b9af-4261-458557f2da50/disk-0.vmdk from lease info. {{(pid=61868) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 954.186075] env[61868]: DEBUG oslo_vmware.rw_handles [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52bef67d-356c-b9af-4261-458557f2da50/disk-0.vmdk for reading. {{(pid=61868) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 954.287638] env[61868]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-a5166d79-efef-4b4e-ab95-a3b6a6a19458 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.304356] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25cf5894-b352-41a6-9d3a-c480400106b3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.320612] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a45506d0-057b-46d0-967b-82659fc1c47f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.354895] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcb1123b-e645-4aec-bed5-fe66ba0092f2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.364121] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5820b0cf-cf99-433b-976c-7c921c763d21 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.379111] env[61868]: DEBUG nova.compute.provider_tree [None req-b062613f-01e0-48de-a1c7-661f529c1378 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 954.389200] env[61868]: DEBUG oslo_vmware.api [None req-1d5307f8-13dc-420d-8f26-1559a2c96086 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315774, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143348} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.390056] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d5307f8-13dc-420d-8f26-1559a2c96086 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 954.390292] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1d5307f8-13dc-420d-8f26-1559a2c96086 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 954.390486] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1d5307f8-13dc-420d-8f26-1559a2c96086 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 954.390666] env[61868]: INFO nova.compute.manager [None req-1d5307f8-13dc-420d-8f26-1559a2c96086 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Took 1.13 seconds to destroy the instance on the hypervisor. [ 954.390915] env[61868]: DEBUG oslo.service.loopingcall [None req-1d5307f8-13dc-420d-8f26-1559a2c96086 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 954.391128] env[61868]: DEBUG nova.compute.manager [-] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 954.391227] env[61868]: DEBUG nova.network.neutron [-] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 954.471604] env[61868]: DEBUG nova.network.neutron [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Successfully created port: 0f66decc-6560-48cd-8015-0c83c8315e10 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 954.593359] env[61868]: DEBUG nova.compute.manager [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 954.674678] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Creating linked-clone VM from snapshot {{(pid=61868) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 954.675340] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-566457c0-91bd-4c0b-b1c9-d10b9cfaa3a8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.685620] env[61868]: DEBUG oslo_vmware.api [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 954.685620] env[61868]: value = "task-1315775" [ 954.685620] env[61868]: _type = "Task" [ 954.685620] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.695507] env[61868]: DEBUG oslo_vmware.api [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315775, 'name': CloneVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.856736] env[61868]: DEBUG nova.compute.manager [req-2bf56a39-667f-494c-a993-59b14b777767 req-eaf3ff69-c584-44a6-b498-5da1f1701805 service nova] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Received event network-vif-deleted-9c333eb1-c59f-4cf4-b25d-2f7a4c6c0a5e {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 954.856956] env[61868]: INFO nova.compute.manager [req-2bf56a39-667f-494c-a993-59b14b777767 req-eaf3ff69-c584-44a6-b498-5da1f1701805 service nova] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Neutron deleted interface 9c333eb1-c59f-4cf4-b25d-2f7a4c6c0a5e; detaching it from the instance and deleting it from the info cache [ 954.857238] env[61868]: DEBUG nova.network.neutron [req-2bf56a39-667f-494c-a993-59b14b777767 req-eaf3ff69-c584-44a6-b498-5da1f1701805 service nova] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.885402] env[61868]: DEBUG nova.scheduler.client.report [None req-b062613f-01e0-48de-a1c7-661f529c1378 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 955.199711] env[61868]: DEBUG oslo_vmware.api [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315775, 'name': CloneVM_Task} progress is 94%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.326562] env[61868]: DEBUG nova.network.neutron [-] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.360775] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f3e9acb2-4162-4f75-9df3-a38ca1c17e13 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.371377] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb207e37-3dfd-440f-8fca-4e278f56661e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.391116] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b062613f-01e0-48de-a1c7-661f529c1378 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.813s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.393760] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e22cc326-cf74-4114-bbe4-7c54f0664481 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 2.686s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.402177] env[61868]: DEBUG nova.compute.manager [req-2bf56a39-667f-494c-a993-59b14b777767 req-eaf3ff69-c584-44a6-b498-5da1f1701805 service nova] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Detach interface failed, port_id=9c333eb1-c59f-4cf4-b25d-2f7a4c6c0a5e, reason: Instance cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b could not be found. {{(pid=61868) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 955.414600] env[61868]: INFO nova.scheduler.client.report [None req-b062613f-01e0-48de-a1c7-661f529c1378 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Deleted allocations for instance 87b25494-1853-4c7b-ae8b-6283992e895c [ 955.603986] env[61868]: DEBUG nova.compute.manager [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 955.630686] env[61868]: DEBUG nova.virt.hardware [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 955.632284] env[61868]: DEBUG nova.virt.hardware [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 955.632284] env[61868]: DEBUG nova.virt.hardware [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 955.632284] env[61868]: DEBUG nova.virt.hardware [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 955.632284] env[61868]: DEBUG nova.virt.hardware [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 955.632284] env[61868]: DEBUG nova.virt.hardware [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 955.632284] env[61868]: DEBUG nova.virt.hardware [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 955.632546] env[61868]: DEBUG nova.virt.hardware [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 955.632695] env[61868]: DEBUG nova.virt.hardware [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 955.633205] env[61868]: DEBUG nova.virt.hardware [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 955.633205] env[61868]: DEBUG nova.virt.hardware [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 955.634128] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c553f924-a430-4e8b-b23a-618a3ffd831f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.643245] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef8f0d85-d4e3-472c-bf43-dfd1ff2c3de3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.697180] env[61868]: DEBUG oslo_vmware.api [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315775, 'name': CloneVM_Task} progress is 94%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.829739] env[61868]: INFO nova.compute.manager [-] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Took 1.44 seconds to deallocate network for instance. [ 955.926369] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b062613f-01e0-48de-a1c7-661f529c1378 tempest-ServerMetadataNegativeTestJSON-974651877 tempest-ServerMetadataNegativeTestJSON-974651877-project-member] Lock "87b25494-1853-4c7b-ae8b-6283992e895c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.647s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.055232] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-637ce0b6-1a88-4f78-8857-564ccee2aaf9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.067160] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cfbc394-fe39-477a-bb6b-29823a41d207 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.116885] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-397f5824-acb4-4277-8e46-db584aa4111c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.125473] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55217df4-4f3c-4ab8-a30f-16db78ea4412 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.146203] env[61868]: DEBUG nova.compute.provider_tree [None req-e22cc326-cf74-4114-bbe4-7c54f0664481 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 956.197544] env[61868]: DEBUG oslo_vmware.api [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315775, 'name': CloneVM_Task, 'duration_secs': 1.486844} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.197972] env[61868]: INFO nova.virt.vmwareapi.vmops [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Created linked-clone VM from snapshot [ 956.198750] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e1e2006-2b34-4033-a45b-44a288e51fba {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.206848] env[61868]: DEBUG nova.virt.vmwareapi.images [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Uploading image f2f4e209-f4e2-4828-a1a0-500857c8d4db {{(pid=61868) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 956.233019] env[61868]: DEBUG oslo_vmware.rw_handles [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 956.233019] env[61868]: value = "vm-281593" [ 956.233019] env[61868]: _type = "VirtualMachine" [ 956.233019] env[61868]: }. {{(pid=61868) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 956.233019] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-f8eebb44-5d27-42cf-8a4d-a9ca5b3c58a2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.243863] env[61868]: DEBUG oslo_vmware.rw_handles [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lease: (returnval){ [ 956.243863] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52b39895-21b5-ca81-c575-c38d7908182c" [ 956.243863] env[61868]: _type = "HttpNfcLease" [ 956.243863] env[61868]: } obtained for exporting VM: (result){ [ 956.243863] env[61868]: value = "vm-281593" [ 956.243863] env[61868]: _type = "VirtualMachine" [ 956.243863] env[61868]: }. {{(pid=61868) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 956.243863] env[61868]: DEBUG oslo_vmware.api [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the lease: (returnval){ [ 956.243863] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52b39895-21b5-ca81-c575-c38d7908182c" [ 956.243863] env[61868]: _type = "HttpNfcLease" [ 956.243863] env[61868]: } to be ready. {{(pid=61868) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 956.251407] env[61868]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 956.251407] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52b39895-21b5-ca81-c575-c38d7908182c" [ 956.251407] env[61868]: _type = "HttpNfcLease" [ 956.251407] env[61868]: } is initializing. {{(pid=61868) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 956.336729] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d5307f8-13dc-420d-8f26-1559a2c96086 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.431328] env[61868]: DEBUG nova.network.neutron [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Successfully updated port: 0f66decc-6560-48cd-8015-0c83c8315e10 {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 956.650200] env[61868]: DEBUG nova.scheduler.client.report [None req-e22cc326-cf74-4114-bbe4-7c54f0664481 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 956.752317] env[61868]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 956.752317] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52b39895-21b5-ca81-c575-c38d7908182c" [ 956.752317] env[61868]: _type = "HttpNfcLease" [ 956.752317] env[61868]: } is ready. {{(pid=61868) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 956.752317] env[61868]: DEBUG oslo_vmware.rw_handles [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 956.752317] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52b39895-21b5-ca81-c575-c38d7908182c" [ 956.752317] env[61868]: _type = "HttpNfcLease" [ 956.752317] env[61868]: }. {{(pid=61868) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 956.752317] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bed13675-cab9-4995-bb8d-0dbf29723250 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.761571] env[61868]: DEBUG oslo_vmware.rw_handles [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522df802-ba0d-9847-d91b-8540e075eddf/disk-0.vmdk from lease info. {{(pid=61868) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 956.761778] env[61868]: DEBUG oslo_vmware.rw_handles [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522df802-ba0d-9847-d91b-8540e075eddf/disk-0.vmdk for reading. {{(pid=61868) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 956.875166] env[61868]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-915460ab-4b0f-4c05-8b7d-7b90d6f997e5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.888487] env[61868]: DEBUG nova.compute.manager [req-7dedc666-c831-4142-a2e4-5d9fcb0fcec2 req-3cdf7357-e4f0-496d-a6bb-d28ad629a5fb service nova] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Received event network-vif-plugged-0f66decc-6560-48cd-8015-0c83c8315e10 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 956.889132] env[61868]: DEBUG oslo_concurrency.lockutils [req-7dedc666-c831-4142-a2e4-5d9fcb0fcec2 req-3cdf7357-e4f0-496d-a6bb-d28ad629a5fb service nova] Acquiring lock "4216e9b6-7b86-409e-9543-1c57e0f4b850-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.889954] env[61868]: DEBUG oslo_concurrency.lockutils [req-7dedc666-c831-4142-a2e4-5d9fcb0fcec2 req-3cdf7357-e4f0-496d-a6bb-d28ad629a5fb service nova] Lock "4216e9b6-7b86-409e-9543-1c57e0f4b850-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.890514] env[61868]: DEBUG oslo_concurrency.lockutils [req-7dedc666-c831-4142-a2e4-5d9fcb0fcec2 req-3cdf7357-e4f0-496d-a6bb-d28ad629a5fb service nova] Lock "4216e9b6-7b86-409e-9543-1c57e0f4b850-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.890514] env[61868]: DEBUG nova.compute.manager [req-7dedc666-c831-4142-a2e4-5d9fcb0fcec2 req-3cdf7357-e4f0-496d-a6bb-d28ad629a5fb service nova] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] No waiting events found dispatching network-vif-plugged-0f66decc-6560-48cd-8015-0c83c8315e10 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 956.890910] env[61868]: WARNING nova.compute.manager [req-7dedc666-c831-4142-a2e4-5d9fcb0fcec2 req-3cdf7357-e4f0-496d-a6bb-d28ad629a5fb service nova] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Received unexpected event network-vif-plugged-0f66decc-6560-48cd-8015-0c83c8315e10 for instance with vm_state building and task_state spawning. [ 956.890910] env[61868]: DEBUG nova.compute.manager [req-7dedc666-c831-4142-a2e4-5d9fcb0fcec2 req-3cdf7357-e4f0-496d-a6bb-d28ad629a5fb service nova] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Received event network-changed-0f66decc-6560-48cd-8015-0c83c8315e10 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 956.890985] env[61868]: DEBUG nova.compute.manager [req-7dedc666-c831-4142-a2e4-5d9fcb0fcec2 req-3cdf7357-e4f0-496d-a6bb-d28ad629a5fb service nova] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Refreshing instance network info cache due to event network-changed-0f66decc-6560-48cd-8015-0c83c8315e10. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 956.891359] env[61868]: DEBUG oslo_concurrency.lockutils [req-7dedc666-c831-4142-a2e4-5d9fcb0fcec2 req-3cdf7357-e4f0-496d-a6bb-d28ad629a5fb service nova] Acquiring lock "refresh_cache-4216e9b6-7b86-409e-9543-1c57e0f4b850" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.891359] env[61868]: DEBUG oslo_concurrency.lockutils [req-7dedc666-c831-4142-a2e4-5d9fcb0fcec2 req-3cdf7357-e4f0-496d-a6bb-d28ad629a5fb service nova] Acquired lock "refresh_cache-4216e9b6-7b86-409e-9543-1c57e0f4b850" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.891512] env[61868]: DEBUG nova.network.neutron [req-7dedc666-c831-4142-a2e4-5d9fcb0fcec2 req-3cdf7357-e4f0-496d-a6bb-d28ad629a5fb service nova] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Refreshing network info cache for port 0f66decc-6560-48cd-8015-0c83c8315e10 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 956.935008] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "refresh_cache-4216e9b6-7b86-409e-9543-1c57e0f4b850" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.442236] env[61868]: DEBUG nova.network.neutron [req-7dedc666-c831-4142-a2e4-5d9fcb0fcec2 req-3cdf7357-e4f0-496d-a6bb-d28ad629a5fb service nova] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 957.550550] env[61868]: DEBUG nova.network.neutron [req-7dedc666-c831-4142-a2e4-5d9fcb0fcec2 req-3cdf7357-e4f0-496d-a6bb-d28ad629a5fb service nova] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.662769] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e22cc326-cf74-4114-bbe4-7c54f0664481 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.269s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.665869] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f65c85db-097a-4399-ba19-932a22223218 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.771s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.666278] env[61868]: DEBUG nova.objects.instance [None req-f65c85db-097a-4399-ba19-932a22223218 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lazy-loading 'resources' on Instance uuid 5aec2609-79d3-4725-a182-70b575adbe17 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 958.060553] env[61868]: DEBUG oslo_concurrency.lockutils [req-7dedc666-c831-4142-a2e4-5d9fcb0fcec2 req-3cdf7357-e4f0-496d-a6bb-d28ad629a5fb service nova] Releasing lock "refresh_cache-4216e9b6-7b86-409e-9543-1c57e0f4b850" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.060553] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquired lock "refresh_cache-4216e9b6-7b86-409e-9543-1c57e0f4b850" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.060553] env[61868]: DEBUG nova.network.neutron [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 958.230190] env[61868]: INFO nova.scheduler.client.report [None req-e22cc326-cf74-4114-bbe4-7c54f0664481 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Deleted allocation for migration 70dd0d99-e65a-47fd-872a-4c6b316b0d07 [ 958.296302] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ce8af3b-4141-408e-bfe7-5b206b94f795 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.305697] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28373379-b3c5-4c8c-8b2d-37c33a41abed {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.344970] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8712452-e0a5-4923-8652-788e6a52ced2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.355979] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d709f42-0a1f-4e55-bbf8-2c3aeb50c941 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.373496] env[61868]: DEBUG nova.compute.provider_tree [None req-f65c85db-097a-4399-ba19-932a22223218 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 958.598225] env[61868]: DEBUG nova.network.neutron [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 958.736290] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e22cc326-cf74-4114-bbe4-7c54f0664481 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "842c8606-632a-47d4-9a67-38d8e6ce6b4d" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 9.120s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.767157] env[61868]: DEBUG nova.network.neutron [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Updating instance_info_cache with network_info: [{"id": "0f66decc-6560-48cd-8015-0c83c8315e10", "address": "fa:16:3e:ee:81:54", "network": {"id": "5b4c5906-b873-4369-855b-e0898ad84812", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-254512637-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "812e115cc1114d0e8536924c4f5556f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0f66decc-65", "ovs_interfaceid": "0f66decc-6560-48cd-8015-0c83c8315e10", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.876723] env[61868]: DEBUG nova.scheduler.client.report [None req-f65c85db-097a-4399-ba19-932a22223218 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 959.270311] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Releasing lock "refresh_cache-4216e9b6-7b86-409e-9543-1c57e0f4b850" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.270709] env[61868]: DEBUG nova.compute.manager [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Instance network_info: |[{"id": "0f66decc-6560-48cd-8015-0c83c8315e10", "address": "fa:16:3e:ee:81:54", "network": {"id": "5b4c5906-b873-4369-855b-e0898ad84812", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-254512637-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "812e115cc1114d0e8536924c4f5556f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0f66decc-65", "ovs_interfaceid": "0f66decc-6560-48cd-8015-0c83c8315e10", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 959.271257] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ee:81:54', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a92a4ffe-7939-4697-bf98-5b22e2c7feda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0f66decc-6560-48cd-8015-0c83c8315e10', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 959.280639] env[61868]: DEBUG oslo.service.loopingcall [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 959.280934] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 959.281325] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8b9cb873-9022-4fad-affe-9de2c901cb40 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.305987] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 959.305987] env[61868]: value = "task-1315777" [ 959.305987] env[61868]: _type = "Task" [ 959.305987] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.315828] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315777, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.383046] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f65c85db-097a-4399-ba19-932a22223218 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.717s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.385677] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d5307f8-13dc-420d-8f26-1559a2c96086 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.049s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.385967] env[61868]: DEBUG nova.objects.instance [None req-1d5307f8-13dc-420d-8f26-1559a2c96086 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lazy-loading 'resources' on Instance uuid cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 959.410638] env[61868]: INFO nova.scheduler.client.report [None req-f65c85db-097a-4399-ba19-932a22223218 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Deleted allocations for instance 5aec2609-79d3-4725-a182-70b575adbe17 [ 959.759481] env[61868]: DEBUG oslo_concurrency.lockutils [None req-97701de6-8f1c-4389-a652-89669d4eb48d tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "e59a182d-97b2-454a-bc40-8afb0839324f" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.759804] env[61868]: DEBUG oslo_concurrency.lockutils [None req-97701de6-8f1c-4389-a652-89669d4eb48d tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "e59a182d-97b2-454a-bc40-8afb0839324f" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.816688] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315777, 'name': CreateVM_Task, 'duration_secs': 0.380311} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.816861] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 959.817562] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.817741] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.818119] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 959.818396] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3b0230db-b410-4b2f-bd92-c5eac6190d96 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.823950] env[61868]: DEBUG oslo_vmware.api [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 959.823950] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52a93f47-f1e2-611a-0706-8685f7c1e265" [ 959.823950] env[61868]: _type = "Task" [ 959.823950] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.833328] env[61868]: DEBUG oslo_vmware.api [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52a93f47-f1e2-611a-0706-8685f7c1e265, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.918866] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f65c85db-097a-4399-ba19-932a22223218 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "5aec2609-79d3-4725-a182-70b575adbe17" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.558s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.919814] env[61868]: DEBUG oslo_concurrency.lockutils [req-0f075bba-e248-47ea-a127-4f37ec996e15 req-8892b0b6-2f80-45f9-8c0d-d5c112c3dbe8 service nova] Acquired lock "5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.920917] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7670628-45cf-4914-958c-f5cd92c9b730 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.934185] env[61868]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 959.934410] env[61868]: DEBUG oslo_vmware.api [-] Fault list: [ManagedObjectNotFound] {{(pid=61868) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 959.934889] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8e3b6143-f658-49fd-956f-adc4d52bb9d6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.948137] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80d53bc1-0af6-4978-b455-afae43e69974 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.985718] env[61868]: ERROR root [req-0f075bba-e248-47ea-a127-4f37ec996e15 req-8892b0b6-2f80-45f9-8c0d-d5c112c3dbe8 service nova] Original exception being dropped: ['Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 377, in request_handler\n response = request(managed_object, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 586, in __call__\n return client.invoke(args, kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 728, in invoke\n result = self.send(soapenv, timeout=timeout)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 777, in send\n return self.process_reply(reply.message, None, None)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 840, in process_reply\n raise WebFault(fault, replyroot)\n', "suds.WebFault: Server raised fault: 'The object 'vim.VirtualMachine:vm-281547' has already been deleted or has not been completely created'\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 301, in _invoke_api\n return api_method(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 480, in get_object_property\n props = get_object_properties(vim, moref, [property_name],\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 360, in get_object_properties\n retrieve_result = vim.RetrievePropertiesEx(\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 413, in request_handler\n raise exceptions.VimFaultException(fault_list, fault_string,\n', "oslo_vmware.exceptions.VimFaultException: The object 'vim.VirtualMachine:vm-281547' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-281547' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-281547'}\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 123, in _call_method\n return self.invoke_api(module, method, self.vim, *args,\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 358, in invoke_api\n return _invoke_api(module, method, *args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 122, in func\n return evt.wait()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait\n result = hub.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch\n return self.greenlet.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 122, in _inner\n idle = self.f(*self.args, **self.kw)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 96, in _func\n result = f(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 341, in _invoke_api\n raise clazz(str(excep),\n', "oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.VirtualMachine:vm-281547' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-281547' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-281547'}\n"]: nova.exception.InstanceNotFound: Instance 5aec2609-79d3-4725-a182-70b575adbe17 could not be found. [ 959.985945] env[61868]: DEBUG oslo_concurrency.lockutils [req-0f075bba-e248-47ea-a127-4f37ec996e15 req-8892b0b6-2f80-45f9-8c0d-d5c112c3dbe8 service nova] Releasing lock "5aec2609-79d3-4725-a182-70b575adbe17" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.986239] env[61868]: DEBUG nova.compute.manager [req-0f075bba-e248-47ea-a127-4f37ec996e15 req-8892b0b6-2f80-45f9-8c0d-d5c112c3dbe8 service nova] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Detach interface failed, port_id=87455df2-50af-4a63-b58b-4813a20d0b38, reason: Instance 5aec2609-79d3-4725-a182-70b575adbe17 could not be found. {{(pid=61868) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 960.007648] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b620a6b-5551-4508-a183-3f2bb3ac100b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.018300] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efb957be-fd45-4531-97b8-a1e5deabb9e6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.051489] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b17289c6-7780-4be5-8c01-ef2a62d6306d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.060188] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b735f7f-4a1a-4a1f-a64b-397ec21a0fcb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.075200] env[61868]: DEBUG nova.compute.provider_tree [None req-1d5307f8-13dc-420d-8f26-1559a2c96086 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 960.263356] env[61868]: DEBUG nova.compute.utils [None req-97701de6-8f1c-4389-a652-89669d4eb48d tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 960.336190] env[61868]: DEBUG oslo_vmware.api [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52a93f47-f1e2-611a-0706-8685f7c1e265, 'name': SearchDatastore_Task, 'duration_secs': 0.012088} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.336556] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.336794] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 960.337079] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.337240] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.337437] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 960.337719] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f6ea4184-231a-4779-9682-dc126feeefea {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.349591] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 960.349815] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 960.350670] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c5039cb-d87b-4053-b8f9-a691a9fc77a3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.357081] env[61868]: DEBUG oslo_vmware.api [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 960.357081] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5289a51d-bb04-917d-4d0e-298d3feaf1dd" [ 960.357081] env[61868]: _type = "Task" [ 960.357081] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.368589] env[61868]: DEBUG oslo_vmware.api [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5289a51d-bb04-917d-4d0e-298d3feaf1dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.578725] env[61868]: DEBUG nova.scheduler.client.report [None req-1d5307f8-13dc-420d-8f26-1559a2c96086 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 960.766963] env[61868]: DEBUG oslo_concurrency.lockutils [None req-97701de6-8f1c-4389-a652-89669d4eb48d tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "e59a182d-97b2-454a-bc40-8afb0839324f" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.869979] env[61868]: DEBUG oslo_vmware.api [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5289a51d-bb04-917d-4d0e-298d3feaf1dd, 'name': SearchDatastore_Task, 'duration_secs': 0.014818} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.870871] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a86291d7-f420-4043-8e28-8414adb412e1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.877393] env[61868]: DEBUG oslo_vmware.api [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 960.877393] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52bfb86c-0443-acad-5be8-3ed670ca5979" [ 960.877393] env[61868]: _type = "Task" [ 960.877393] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.886959] env[61868]: DEBUG oslo_vmware.api [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52bfb86c-0443-acad-5be8-3ed670ca5979, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.084501] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d5307f8-13dc-420d-8f26-1559a2c96086 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.699s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.108295] env[61868]: INFO nova.scheduler.client.report [None req-1d5307f8-13dc-420d-8f26-1559a2c96086 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Deleted allocations for instance cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b [ 961.390387] env[61868]: DEBUG oslo_vmware.api [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52bfb86c-0443-acad-5be8-3ed670ca5979, 'name': SearchDatastore_Task, 'duration_secs': 0.015083} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.390754] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.391162] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 4216e9b6-7b86-409e-9543-1c57e0f4b850/4216e9b6-7b86-409e-9543-1c57e0f4b850.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 961.391538] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7b6f3d4e-00e6-460f-be5f-1431f3759e2c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.401169] env[61868]: DEBUG oslo_vmware.api [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 961.401169] env[61868]: value = "task-1315778" [ 961.401169] env[61868]: _type = "Task" [ 961.401169] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.410785] env[61868]: DEBUG oslo_vmware.api [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315778, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.617531] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1d5307f8-13dc-420d-8f26-1559a2c96086 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.366s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.669551] env[61868]: DEBUG oslo_vmware.rw_handles [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52bef67d-356c-b9af-4261-458557f2da50/disk-0.vmdk. {{(pid=61868) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 961.671470] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1bf7550-8eee-4eb2-b504-cb43ead43b76 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.681659] env[61868]: DEBUG oslo_vmware.rw_handles [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52bef67d-356c-b9af-4261-458557f2da50/disk-0.vmdk is in state: ready. {{(pid=61868) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 961.682044] env[61868]: ERROR oslo_vmware.rw_handles [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52bef67d-356c-b9af-4261-458557f2da50/disk-0.vmdk due to incomplete transfer. [ 961.682439] env[61868]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-c362e9e2-29d2-4549-937f-9559a0a2e72c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.693224] env[61868]: DEBUG oslo_vmware.rw_handles [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52bef67d-356c-b9af-4261-458557f2da50/disk-0.vmdk. {{(pid=61868) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 961.693694] env[61868]: DEBUG nova.virt.vmwareapi.images [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Uploaded image 76af0b0a-47a5-487a-9b10-d16668b6caae to the Glance image server {{(pid=61868) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 961.699318] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Destroying the VM {{(pid=61868) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 961.699794] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-54775529-5b32-4661-bd7d-0df4d634e774 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.710214] env[61868]: DEBUG oslo_vmware.api [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the task: (returnval){ [ 961.710214] env[61868]: value = "task-1315779" [ 961.710214] env[61868]: _type = "Task" [ 961.710214] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.724122] env[61868]: DEBUG oslo_vmware.api [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315779, 'name': Destroy_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.842905] env[61868]: DEBUG oslo_concurrency.lockutils [None req-97701de6-8f1c-4389-a652-89669d4eb48d tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "e59a182d-97b2-454a-bc40-8afb0839324f" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.843356] env[61868]: DEBUG oslo_concurrency.lockutils [None req-97701de6-8f1c-4389-a652-89669d4eb48d tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "e59a182d-97b2-454a-bc40-8afb0839324f" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.843553] env[61868]: INFO nova.compute.manager [None req-97701de6-8f1c-4389-a652-89669d4eb48d tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Attaching volume d18e4cb9-4034-4576-90e4-7458bce2ca23 to /dev/sdb [ 961.886202] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b384300d-0644-4e1f-a738-3ef1725b70e7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.896939] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a955beb0-f02c-43bd-86ab-7c57a8290c52 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.914344] env[61868]: DEBUG oslo_vmware.api [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315778, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.919184] env[61868]: DEBUG nova.virt.block_device [None req-97701de6-8f1c-4389-a652-89669d4eb48d tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Updating existing volume attachment record: c8859f5b-a5b1-40be-9b85-2642ff1dd575 {{(pid=61868) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 962.227275] env[61868]: DEBUG oslo_vmware.api [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315779, 'name': Destroy_Task} progress is 100%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.253275] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7de8b1d5-6af1-44ba-9aaf-67e96fc429b9 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "842c8606-632a-47d4-9a67-38d8e6ce6b4d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.253275] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7de8b1d5-6af1-44ba-9aaf-67e96fc429b9 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "842c8606-632a-47d4-9a67-38d8e6ce6b4d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.253275] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7de8b1d5-6af1-44ba-9aaf-67e96fc429b9 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "842c8606-632a-47d4-9a67-38d8e6ce6b4d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.253275] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7de8b1d5-6af1-44ba-9aaf-67e96fc429b9 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "842c8606-632a-47d4-9a67-38d8e6ce6b4d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.253275] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7de8b1d5-6af1-44ba-9aaf-67e96fc429b9 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "842c8606-632a-47d4-9a67-38d8e6ce6b4d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.253275] env[61868]: INFO nova.compute.manager [None req-7de8b1d5-6af1-44ba-9aaf-67e96fc429b9 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Terminating instance [ 962.259906] env[61868]: DEBUG nova.compute.manager [None req-7de8b1d5-6af1-44ba-9aaf-67e96fc429b9 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 962.259906] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-7de8b1d5-6af1-44ba-9aaf-67e96fc429b9 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 962.259906] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a106e1e6-c98c-4f00-97df-f0d9cf2f0d7b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.269965] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-7de8b1d5-6af1-44ba-9aaf-67e96fc429b9 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 962.269965] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-46bcf53a-5013-4e1c-b1e9-c995f83ab90f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.286224] env[61868]: DEBUG oslo_vmware.api [None req-7de8b1d5-6af1-44ba-9aaf-67e96fc429b9 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 962.286224] env[61868]: value = "task-1315783" [ 962.286224] env[61868]: _type = "Task" [ 962.286224] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.296570] env[61868]: DEBUG oslo_vmware.api [None req-7de8b1d5-6af1-44ba-9aaf-67e96fc429b9 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315783, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.414072] env[61868]: DEBUG oslo_vmware.api [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315778, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.644413} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.414652] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 4216e9b6-7b86-409e-9543-1c57e0f4b850/4216e9b6-7b86-409e-9543-1c57e0f4b850.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 962.414884] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 962.415184] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5a067214-07f6-46b8-a38d-63f2034d0683 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.425434] env[61868]: DEBUG oslo_vmware.api [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 962.425434] env[61868]: value = "task-1315784" [ 962.425434] env[61868]: _type = "Task" [ 962.425434] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.436514] env[61868]: DEBUG oslo_vmware.api [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315784, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.722336] env[61868]: DEBUG oslo_vmware.api [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315779, 'name': Destroy_Task} progress is 100%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.797794] env[61868]: DEBUG oslo_vmware.api [None req-7de8b1d5-6af1-44ba-9aaf-67e96fc429b9 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315783, 'name': PowerOffVM_Task, 'duration_secs': 0.207621} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.798090] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-7de8b1d5-6af1-44ba-9aaf-67e96fc429b9 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 962.798268] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-7de8b1d5-6af1-44ba-9aaf-67e96fc429b9 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 962.798525] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-009ca63c-1a7c-4fb6-83e1-e7625ef1237d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.895437] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-7de8b1d5-6af1-44ba-9aaf-67e96fc429b9 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 962.895741] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-7de8b1d5-6af1-44ba-9aaf-67e96fc429b9 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Deleting contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 962.895862] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-7de8b1d5-6af1-44ba-9aaf-67e96fc429b9 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Deleting the datastore file [datastore2] 842c8606-632a-47d4-9a67-38d8e6ce6b4d {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 962.896183] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-52555ad0-f5dd-4f39-b130-92791da1d8b0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.904934] env[61868]: DEBUG oslo_vmware.api [None req-7de8b1d5-6af1-44ba-9aaf-67e96fc429b9 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 962.904934] env[61868]: value = "task-1315786" [ 962.904934] env[61868]: _type = "Task" [ 962.904934] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.915240] env[61868]: DEBUG oslo_vmware.api [None req-7de8b1d5-6af1-44ba-9aaf-67e96fc429b9 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315786, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.935880] env[61868]: DEBUG oslo_vmware.api [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315784, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.091855} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.936310] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 962.937352] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f27e2331-e679-4239-b96b-1b4d1b957355 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.963716] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] 4216e9b6-7b86-409e-9543-1c57e0f4b850/4216e9b6-7b86-409e-9543-1c57e0f4b850.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 962.964102] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-32a2f988-4c0c-4234-b367-8793ad6b77a0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.986287] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "5d2b56d5-a4e0-4dba-9158-71036d60a40b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.986567] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "5d2b56d5-a4e0-4dba-9158-71036d60a40b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.987719] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "9286cc76-1152-4fd4-9279-33d6ee4b9486" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.987946] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "9286cc76-1152-4fd4-9279-33d6ee4b9486" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.991528] env[61868]: DEBUG oslo_vmware.api [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 962.991528] env[61868]: value = "task-1315787" [ 962.991528] env[61868]: _type = "Task" [ 962.991528] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.001195] env[61868]: DEBUG oslo_vmware.api [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315787, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.223801] env[61868]: DEBUG oslo_vmware.api [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315779, 'name': Destroy_Task, 'duration_secs': 1.171004} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.224098] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Destroyed the VM [ 963.224483] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Deleting Snapshot of the VM instance {{(pid=61868) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 963.224730] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-ca4812bf-7b91-4bfb-a6c1-faa78f159c79 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.232146] env[61868]: DEBUG oslo_vmware.api [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the task: (returnval){ [ 963.232146] env[61868]: value = "task-1315788" [ 963.232146] env[61868]: _type = "Task" [ 963.232146] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.241356] env[61868]: DEBUG oslo_vmware.api [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315788, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.417433] env[61868]: DEBUG oslo_vmware.api [None req-7de8b1d5-6af1-44ba-9aaf-67e96fc429b9 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315786, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.187176} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.418118] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-7de8b1d5-6af1-44ba-9aaf-67e96fc429b9 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 963.418118] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-7de8b1d5-6af1-44ba-9aaf-67e96fc429b9 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Deleted contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 963.418118] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-7de8b1d5-6af1-44ba-9aaf-67e96fc429b9 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 963.418280] env[61868]: INFO nova.compute.manager [None req-7de8b1d5-6af1-44ba-9aaf-67e96fc429b9 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Took 1.16 seconds to destroy the instance on the hypervisor. [ 963.418632] env[61868]: DEBUG oslo.service.loopingcall [None req-7de8b1d5-6af1-44ba-9aaf-67e96fc429b9 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 963.418898] env[61868]: DEBUG nova.compute.manager [-] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 963.422224] env[61868]: DEBUG nova.network.neutron [-] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 963.491850] env[61868]: DEBUG nova.compute.manager [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 963.492778] env[61868]: DEBUG nova.compute.manager [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 963.511021] env[61868]: DEBUG oslo_vmware.api [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315787, 'name': ReconfigVM_Task, 'duration_secs': 0.351544} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.511021] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Reconfigured VM instance instance-00000057 to attach disk [datastore2] 4216e9b6-7b86-409e-9543-1c57e0f4b850/4216e9b6-7b86-409e-9543-1c57e0f4b850.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 963.512953] env[61868]: DEBUG oslo_vmware.rw_handles [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522df802-ba0d-9847-d91b-8540e075eddf/disk-0.vmdk. {{(pid=61868) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 963.513256] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-85f05962-8a5f-42b2-9f5a-1bb046de8916 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.515624] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-655b8aad-009b-4e70-9532-013379108cbf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.525078] env[61868]: DEBUG oslo_vmware.rw_handles [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522df802-ba0d-9847-d91b-8540e075eddf/disk-0.vmdk is in state: ready. {{(pid=61868) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 963.525078] env[61868]: ERROR oslo_vmware.rw_handles [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522df802-ba0d-9847-d91b-8540e075eddf/disk-0.vmdk due to incomplete transfer. [ 963.526341] env[61868]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-e88ccdab-3d9c-4b58-bf79-07667aea2303 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.527965] env[61868]: DEBUG oslo_vmware.api [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 963.527965] env[61868]: value = "task-1315789" [ 963.527965] env[61868]: _type = "Task" [ 963.527965] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.535596] env[61868]: DEBUG oslo_vmware.rw_handles [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522df802-ba0d-9847-d91b-8540e075eddf/disk-0.vmdk. {{(pid=61868) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 963.535811] env[61868]: DEBUG nova.virt.vmwareapi.images [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Uploaded image f2f4e209-f4e2-4828-a1a0-500857c8d4db to the Glance image server {{(pid=61868) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 963.537761] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Destroying the VM {{(pid=61868) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 963.541297] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-0b415194-ee0c-40da-8b74-f4c6979f019b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.543046] env[61868]: DEBUG oslo_vmware.api [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315789, 'name': Rename_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.550822] env[61868]: DEBUG oslo_vmware.api [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 963.550822] env[61868]: value = "task-1315790" [ 963.550822] env[61868]: _type = "Task" [ 963.550822] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.560299] env[61868]: DEBUG oslo_vmware.api [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315790, 'name': Destroy_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.732898] env[61868]: DEBUG nova.compute.manager [req-e29bc2e8-3d6f-4ae2-b7b0-168beb41904b req-8f0cac55-71ce-40bf-8d62-0e5f7afdc757 service nova] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Received event network-vif-deleted-e3943fa9-423d-4168-8aa6-1651280cd127 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 963.733169] env[61868]: INFO nova.compute.manager [req-e29bc2e8-3d6f-4ae2-b7b0-168beb41904b req-8f0cac55-71ce-40bf-8d62-0e5f7afdc757 service nova] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Neutron deleted interface e3943fa9-423d-4168-8aa6-1651280cd127; detaching it from the instance and deleting it from the info cache [ 963.733380] env[61868]: DEBUG nova.network.neutron [req-e29bc2e8-3d6f-4ae2-b7b0-168beb41904b req-8f0cac55-71ce-40bf-8d62-0e5f7afdc757 service nova] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.745370] env[61868]: DEBUG oslo_vmware.api [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315788, 'name': RemoveSnapshot_Task, 'duration_secs': 0.361525} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.746162] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Deleted Snapshot of the VM instance {{(pid=61868) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 963.746488] env[61868]: DEBUG nova.compute.manager [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 963.747956] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-483e9838-73f6-4360-8bef-58fe27d3dac8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.961200] env[61868]: DEBUG oslo_concurrency.lockutils [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "a96cab5a-c053-4e45-96f6-2aba0a819110" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.961566] env[61868]: DEBUG oslo_concurrency.lockutils [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "a96cab5a-c053-4e45-96f6-2aba0a819110" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.017596] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.017940] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.019472] env[61868]: INFO nova.compute.claims [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 964.023602] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.041500] env[61868]: DEBUG oslo_vmware.api [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315789, 'name': Rename_Task, 'duration_secs': 0.157301} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.042210] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 964.042210] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cb792a6b-bf35-47ef-8ea6-68aba23a5148 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.049706] env[61868]: DEBUG oslo_vmware.api [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 964.049706] env[61868]: value = "task-1315791" [ 964.049706] env[61868]: _type = "Task" [ 964.049706] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.061800] env[61868]: DEBUG oslo_vmware.api [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315791, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.066391] env[61868]: DEBUG oslo_vmware.api [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315790, 'name': Destroy_Task, 'duration_secs': 0.316954} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.066640] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Destroyed the VM [ 964.066876] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Deleting Snapshot of the VM instance {{(pid=61868) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 964.067165] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-b70a2222-0dd6-4cac-8db8-9756cc834f36 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.074487] env[61868]: DEBUG oslo_vmware.api [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 964.074487] env[61868]: value = "task-1315792" [ 964.074487] env[61868]: _type = "Task" [ 964.074487] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.084226] env[61868]: DEBUG oslo_vmware.api [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315792, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.212993] env[61868]: DEBUG nova.network.neutron [-] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.236152] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-55c409ca-83f1-4b2c-992a-b2b4b37fcbc5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.247918] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cebfb32c-5474-4409-984e-2ca08c522455 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.262906] env[61868]: INFO nova.compute.manager [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Shelve offloading [ 964.264720] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 964.264968] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6e465ee8-60d8-4eb7-8733-40cfb31d571b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.283058] env[61868]: DEBUG oslo_vmware.api [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the task: (returnval){ [ 964.283058] env[61868]: value = "task-1315793" [ 964.283058] env[61868]: _type = "Task" [ 964.283058] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.293883] env[61868]: DEBUG nova.compute.manager [req-e29bc2e8-3d6f-4ae2-b7b0-168beb41904b req-8f0cac55-71ce-40bf-8d62-0e5f7afdc757 service nova] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Detach interface failed, port_id=e3943fa9-423d-4168-8aa6-1651280cd127, reason: Instance 842c8606-632a-47d4-9a67-38d8e6ce6b4d could not be found. {{(pid=61868) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 964.299681] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] VM already powered off {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 964.299887] env[61868]: DEBUG nova.compute.manager [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 964.300672] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dac250cf-ec6e-4750-b7fa-fb9daf824edf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.306655] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquiring lock "refresh_cache-75fdbc11-1ef3-442d-bcf4-55069ba9ab62" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.306829] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquired lock "refresh_cache-75fdbc11-1ef3-442d-bcf4-55069ba9ab62" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.307011] env[61868]: DEBUG nova.network.neutron [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 964.463843] env[61868]: DEBUG nova.compute.manager [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 964.473803] env[61868]: DEBUG oslo_concurrency.lockutils [None req-39bcf7a5-825e-47cc-9e88-e67d23c1c42e tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "867384c6-175b-4848-b9ac-917470742ba2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.474183] env[61868]: DEBUG oslo_concurrency.lockutils [None req-39bcf7a5-825e-47cc-9e88-e67d23c1c42e tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "867384c6-175b-4848-b9ac-917470742ba2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.474479] env[61868]: DEBUG oslo_concurrency.lockutils [None req-39bcf7a5-825e-47cc-9e88-e67d23c1c42e tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "867384c6-175b-4848-b9ac-917470742ba2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.474737] env[61868]: DEBUG oslo_concurrency.lockutils [None req-39bcf7a5-825e-47cc-9e88-e67d23c1c42e tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "867384c6-175b-4848-b9ac-917470742ba2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.475360] env[61868]: DEBUG oslo_concurrency.lockutils [None req-39bcf7a5-825e-47cc-9e88-e67d23c1c42e tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "867384c6-175b-4848-b9ac-917470742ba2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.477542] env[61868]: INFO nova.compute.manager [None req-39bcf7a5-825e-47cc-9e88-e67d23c1c42e tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Terminating instance [ 964.480013] env[61868]: DEBUG nova.compute.manager [None req-39bcf7a5-825e-47cc-9e88-e67d23c1c42e tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 964.480429] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-39bcf7a5-825e-47cc-9e88-e67d23c1c42e tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 964.481942] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32930fea-b26c-42ba-a31e-69e3de9f26c1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.491261] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-39bcf7a5-825e-47cc-9e88-e67d23c1c42e tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 964.491907] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f8c79aeb-8092-43ee-b8c7-901fad87e424 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.561915] env[61868]: DEBUG oslo_vmware.api [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315791, 'name': PowerOnVM_Task, 'duration_secs': 0.469871} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.562241] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 964.562467] env[61868]: INFO nova.compute.manager [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Took 8.96 seconds to spawn the instance on the hypervisor. [ 964.562671] env[61868]: DEBUG nova.compute.manager [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 964.563502] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d889a5c8-75f0-4402-94f1-e7c5ff6acb62 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.583822] env[61868]: DEBUG oslo_vmware.api [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315792, 'name': RemoveSnapshot_Task, 'duration_secs': 0.445391} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.583822] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Deleted Snapshot of the VM instance {{(pid=61868) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 964.583822] env[61868]: INFO nova.compute.manager [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Took 11.51 seconds to snapshot the instance on the hypervisor. [ 964.612926] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-39bcf7a5-825e-47cc-9e88-e67d23c1c42e tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 964.613312] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-39bcf7a5-825e-47cc-9e88-e67d23c1c42e tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 964.613668] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-39bcf7a5-825e-47cc-9e88-e67d23c1c42e tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Deleting the datastore file [datastore1] 867384c6-175b-4848-b9ac-917470742ba2 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 964.613974] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b9d8b599-91be-42f2-b5db-e5289634fcbd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.622890] env[61868]: DEBUG oslo_vmware.api [None req-39bcf7a5-825e-47cc-9e88-e67d23c1c42e tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 964.622890] env[61868]: value = "task-1315796" [ 964.622890] env[61868]: _type = "Task" [ 964.622890] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.631369] env[61868]: DEBUG oslo_vmware.api [None req-39bcf7a5-825e-47cc-9e88-e67d23c1c42e tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315796, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.715536] env[61868]: INFO nova.compute.manager [-] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Took 1.30 seconds to deallocate network for instance. [ 964.945746] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "cbeeb7b1-166b-49f7-88c5-6ed05c9dc759" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.945983] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "cbeeb7b1-166b-49f7-88c5-6ed05c9dc759" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.983521] env[61868]: DEBUG oslo_concurrency.lockutils [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.022112] env[61868]: DEBUG nova.network.neutron [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Updating instance_info_cache with network_info: [{"id": "21d6cab8-f691-425c-bfa8-9d2f1c294b64", "address": "fa:16:3e:76:a4:70", "network": {"id": "e7457f1a-171a-4c82-a260-1d1a7a95861e", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1032778554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b9b800086764b41a8817d1d8e463a2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21d6cab8-f6", "ovs_interfaceid": "21d6cab8-f691-425c-bfa8-9d2f1c294b64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.083368] env[61868]: INFO nova.compute.manager [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Took 18.63 seconds to build instance. [ 965.087918] env[61868]: DEBUG nova.compute.manager [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Instance disappeared during snapshot {{(pid=61868) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 965.100508] env[61868]: DEBUG nova.compute.manager [None req-4de2eae6-7ac4-4e2d-b3f5-ea6428de6d71 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Image not found during clean up f2f4e209-f4e2-4828-a1a0-500857c8d4db {{(pid=61868) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4500}} [ 965.135973] env[61868]: DEBUG oslo_vmware.api [None req-39bcf7a5-825e-47cc-9e88-e67d23c1c42e tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315796, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.148206] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f355e35d-825c-493d-b721-33b08c12590a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.155785] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed65cef0-132a-49a3-acef-3a67e77addc0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.189490] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2c596ad-deac-4b43-9b16-a116183ae2b9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.201104] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c48a1f1f-3b7a-4638-8e70-91e2996de1de {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.215956] env[61868]: DEBUG nova.compute.provider_tree [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 965.221761] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7de8b1d5-6af1-44ba-9aaf-67e96fc429b9 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.448742] env[61868]: DEBUG nova.compute.manager [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 965.524670] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Releasing lock "refresh_cache-75fdbc11-1ef3-442d-bcf4-55069ba9ab62" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.585399] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0373ff38-dfd4-477f-bc07-451104a57959 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "4216e9b6-7b86-409e-9543-1c57e0f4b850" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.136s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.636932] env[61868]: DEBUG oslo_vmware.api [None req-39bcf7a5-825e-47cc-9e88-e67d23c1c42e tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315796, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.662697} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.637239] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-39bcf7a5-825e-47cc-9e88-e67d23c1c42e tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 965.637388] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-39bcf7a5-825e-47cc-9e88-e67d23c1c42e tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 965.637569] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-39bcf7a5-825e-47cc-9e88-e67d23c1c42e tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 965.637742] env[61868]: INFO nova.compute.manager [None req-39bcf7a5-825e-47cc-9e88-e67d23c1c42e tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Took 1.16 seconds to destroy the instance on the hypervisor. [ 965.637982] env[61868]: DEBUG oslo.service.loopingcall [None req-39bcf7a5-825e-47cc-9e88-e67d23c1c42e tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 965.638195] env[61868]: DEBUG nova.compute.manager [-] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 965.638288] env[61868]: DEBUG nova.network.neutron [-] [instance: 867384c6-175b-4848-b9ac-917470742ba2] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 965.697940] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60d6a83d-5b6a-4036-9102-b4a7e9832012 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.705447] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-891949e8-ec3a-4a14-9bdb-5c804103c2b8 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Suspending the VM {{(pid=61868) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 965.705688] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-8e56ea85-3884-4af5-9d25-bab57d3823dc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.711803] env[61868]: DEBUG oslo_vmware.api [None req-891949e8-ec3a-4a14-9bdb-5c804103c2b8 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 965.711803] env[61868]: value = "task-1315797" [ 965.711803] env[61868]: _type = "Task" [ 965.711803] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.720926] env[61868]: DEBUG nova.scheduler.client.report [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 965.724569] env[61868]: DEBUG oslo_vmware.api [None req-891949e8-ec3a-4a14-9bdb-5c804103c2b8 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315797, 'name': SuspendVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.905085] env[61868]: DEBUG nova.compute.manager [req-2d10b66d-f095-4e06-a169-22c020e89ced req-11ded042-a1f5-41c4-a8cb-21eed1efe0c1 service nova] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Received event network-vif-unplugged-21d6cab8-f691-425c-bfa8-9d2f1c294b64 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 965.905345] env[61868]: DEBUG oslo_concurrency.lockutils [req-2d10b66d-f095-4e06-a169-22c020e89ced req-11ded042-a1f5-41c4-a8cb-21eed1efe0c1 service nova] Acquiring lock "75fdbc11-1ef3-442d-bcf4-55069ba9ab62-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.905534] env[61868]: DEBUG oslo_concurrency.lockutils [req-2d10b66d-f095-4e06-a169-22c020e89ced req-11ded042-a1f5-41c4-a8cb-21eed1efe0c1 service nova] Lock "75fdbc11-1ef3-442d-bcf4-55069ba9ab62-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.905713] env[61868]: DEBUG oslo_concurrency.lockutils [req-2d10b66d-f095-4e06-a169-22c020e89ced req-11ded042-a1f5-41c4-a8cb-21eed1efe0c1 service nova] Lock "75fdbc11-1ef3-442d-bcf4-55069ba9ab62-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.905902] env[61868]: DEBUG nova.compute.manager [req-2d10b66d-f095-4e06-a169-22c020e89ced req-11ded042-a1f5-41c4-a8cb-21eed1efe0c1 service nova] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] No waiting events found dispatching network-vif-unplugged-21d6cab8-f691-425c-bfa8-9d2f1c294b64 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 965.906126] env[61868]: WARNING nova.compute.manager [req-2d10b66d-f095-4e06-a169-22c020e89ced req-11ded042-a1f5-41c4-a8cb-21eed1efe0c1 service nova] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Received unexpected event network-vif-unplugged-21d6cab8-f691-425c-bfa8-9d2f1c294b64 for instance with vm_state shelved and task_state shelving_offloading. [ 965.958460] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 965.959398] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97f6ab3a-46b1-426d-b5bf-c663cc787ab6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.968847] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 965.969259] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8089744f-4f9e-471a-a647-e109f7f5b8d9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.977186] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.080720] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 966.081079] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 966.081164] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Deleting the datastore file [datastore1] 75fdbc11-1ef3-442d-bcf4-55069ba9ab62 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 966.081455] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e0395db9-46e6-46c9-a39a-16941728f45b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.089163] env[61868]: DEBUG oslo_vmware.api [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the task: (returnval){ [ 966.089163] env[61868]: value = "task-1315799" [ 966.089163] env[61868]: _type = "Task" [ 966.089163] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.097245] env[61868]: DEBUG oslo_vmware.api [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315799, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.224818] env[61868]: DEBUG oslo_vmware.api [None req-891949e8-ec3a-4a14-9bdb-5c804103c2b8 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315797, 'name': SuspendVM_Task} progress is 70%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.226641] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.209s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.227161] env[61868]: DEBUG nova.compute.manager [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 966.229807] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.207s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.231265] env[61868]: INFO nova.compute.claims [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 966.477195] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-97701de6-8f1c-4389-a652-89669d4eb48d tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Volume attach. Driver type: vmdk {{(pid=61868) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 966.477513] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-97701de6-8f1c-4389-a652-89669d4eb48d tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281596', 'volume_id': 'd18e4cb9-4034-4576-90e4-7458bce2ca23', 'name': 'volume-d18e4cb9-4034-4576-90e4-7458bce2ca23', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e59a182d-97b2-454a-bc40-8afb0839324f', 'attached_at': '', 'detached_at': '', 'volume_id': 'd18e4cb9-4034-4576-90e4-7458bce2ca23', 'serial': 'd18e4cb9-4034-4576-90e4-7458bce2ca23'} {{(pid=61868) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 966.478434] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e94531ef-ab0b-4267-9887-1d48a14b9c62 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.496628] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02971493-2454-41a5-a534-e9a7a0e65218 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.522353] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-97701de6-8f1c-4389-a652-89669d4eb48d tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] volume-d18e4cb9-4034-4576-90e4-7458bce2ca23/volume-d18e4cb9-4034-4576-90e4-7458bce2ca23.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 966.522666] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-52d18635-12fd-4caf-bb2a-f10169e5184b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.535396] env[61868]: DEBUG nova.network.neutron [-] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.542628] env[61868]: DEBUG oslo_vmware.api [None req-97701de6-8f1c-4389-a652-89669d4eb48d tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 966.542628] env[61868]: value = "task-1315800" [ 966.542628] env[61868]: _type = "Task" [ 966.542628] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.551363] env[61868]: DEBUG oslo_vmware.api [None req-97701de6-8f1c-4389-a652-89669d4eb48d tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315800, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.600446] env[61868]: DEBUG oslo_vmware.api [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315799, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.157219} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.600719] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 966.600928] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 966.601138] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 966.622204] env[61868]: INFO nova.scheduler.client.report [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Deleted allocations for instance 75fdbc11-1ef3-442d-bcf4-55069ba9ab62 [ 966.724476] env[61868]: DEBUG oslo_vmware.api [None req-891949e8-ec3a-4a14-9bdb-5c804103c2b8 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315797, 'name': SuspendVM_Task, 'duration_secs': 0.625652} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.724637] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-891949e8-ec3a-4a14-9bdb-5c804103c2b8 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Suspended the VM {{(pid=61868) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 966.724930] env[61868]: DEBUG nova.compute.manager [None req-891949e8-ec3a-4a14-9bdb-5c804103c2b8 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 966.725806] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af496c3c-597f-4ad5-a513-54a7f10e6109 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.735291] env[61868]: DEBUG nova.compute.utils [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 966.738295] env[61868]: DEBUG nova.compute.manager [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 966.738450] env[61868]: DEBUG nova.network.neutron [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 966.786832] env[61868]: DEBUG nova.policy [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'daad105e7edf4fb0ae0b2e685bfd92e9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b520c3ac58074e8d9b0bfafb817244a8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 967.039509] env[61868]: INFO nova.compute.manager [-] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Took 1.40 seconds to deallocate network for instance. [ 967.043191] env[61868]: DEBUG nova.network.neutron [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Successfully created port: 4edbd65b-1856-47a2-95c9-e4c8f4d04c71 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 967.055135] env[61868]: DEBUG oslo_vmware.api [None req-97701de6-8f1c-4389-a652-89669d4eb48d tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315800, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.127036] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.239596] env[61868]: DEBUG nova.compute.manager [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 967.363759] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-875e2224-ec4a-46b5-a98b-9efe107a81bf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.372423] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beda882c-da39-4f9b-8813-2a63cd599cf4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.403634] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dc2a192-0170-4cb0-bb64-84fd2ad065c5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.411848] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f105d15e-9ace-4c87-995c-e6a12dcfed6c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.426155] env[61868]: DEBUG nova.compute.provider_tree [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 967.546474] env[61868]: DEBUG oslo_concurrency.lockutils [None req-39bcf7a5-825e-47cc-9e88-e67d23c1c42e tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.556575] env[61868]: DEBUG oslo_vmware.api [None req-97701de6-8f1c-4389-a652-89669d4eb48d tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315800, 'name': ReconfigVM_Task, 'duration_secs': 0.601669} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.556871] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-97701de6-8f1c-4389-a652-89669d4eb48d tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Reconfigured VM instance instance-0000004f to attach disk [datastore2] volume-d18e4cb9-4034-4576-90e4-7458bce2ca23/volume-d18e4cb9-4034-4576-90e4-7458bce2ca23.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 967.561641] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-56371f80-6c67-4825-a495-aa192df2046b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.578256] env[61868]: DEBUG oslo_vmware.api [None req-97701de6-8f1c-4389-a652-89669d4eb48d tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 967.578256] env[61868]: value = "task-1315801" [ 967.578256] env[61868]: _type = "Task" [ 967.578256] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.587320] env[61868]: DEBUG oslo_vmware.api [None req-97701de6-8f1c-4389-a652-89669d4eb48d tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315801, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.929571] env[61868]: DEBUG nova.scheduler.client.report [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 967.934835] env[61868]: DEBUG nova.compute.manager [req-611eb4b8-4feb-4851-876e-90189d2eaf05 req-4cc4ea63-4460-4add-b249-288cf6bd9a1f service nova] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Received event network-changed-21d6cab8-f691-425c-bfa8-9d2f1c294b64 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 967.935048] env[61868]: DEBUG nova.compute.manager [req-611eb4b8-4feb-4851-876e-90189d2eaf05 req-4cc4ea63-4460-4add-b249-288cf6bd9a1f service nova] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Refreshing instance network info cache due to event network-changed-21d6cab8-f691-425c-bfa8-9d2f1c294b64. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 967.935274] env[61868]: DEBUG oslo_concurrency.lockutils [req-611eb4b8-4feb-4851-876e-90189d2eaf05 req-4cc4ea63-4460-4add-b249-288cf6bd9a1f service nova] Acquiring lock "refresh_cache-75fdbc11-1ef3-442d-bcf4-55069ba9ab62" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.935418] env[61868]: DEBUG oslo_concurrency.lockutils [req-611eb4b8-4feb-4851-876e-90189d2eaf05 req-4cc4ea63-4460-4add-b249-288cf6bd9a1f service nova] Acquired lock "refresh_cache-75fdbc11-1ef3-442d-bcf4-55069ba9ab62" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.935585] env[61868]: DEBUG nova.network.neutron [req-611eb4b8-4feb-4851-876e-90189d2eaf05 req-4cc4ea63-4460-4add-b249-288cf6bd9a1f service nova] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Refreshing network info cache for port 21d6cab8-f691-425c-bfa8-9d2f1c294b64 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 968.089273] env[61868]: DEBUG oslo_vmware.api [None req-97701de6-8f1c-4389-a652-89669d4eb48d tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315801, 'name': ReconfigVM_Task, 'duration_secs': 0.14198} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.089573] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-97701de6-8f1c-4389-a652-89669d4eb48d tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281596', 'volume_id': 'd18e4cb9-4034-4576-90e4-7458bce2ca23', 'name': 'volume-d18e4cb9-4034-4576-90e4-7458bce2ca23', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e59a182d-97b2-454a-bc40-8afb0839324f', 'attached_at': '', 'detached_at': '', 'volume_id': 'd18e4cb9-4034-4576-90e4-7458bce2ca23', 'serial': 'd18e4cb9-4034-4576-90e4-7458bce2ca23'} {{(pid=61868) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 968.125203] env[61868]: DEBUG oslo_concurrency.lockutils [None req-801169df-e652-4a2f-9396-52c991a74fc1 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "4216e9b6-7b86-409e-9543-1c57e0f4b850" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.125438] env[61868]: DEBUG oslo_concurrency.lockutils [None req-801169df-e652-4a2f-9396-52c991a74fc1 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "4216e9b6-7b86-409e-9543-1c57e0f4b850" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.125644] env[61868]: DEBUG oslo_concurrency.lockutils [None req-801169df-e652-4a2f-9396-52c991a74fc1 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "4216e9b6-7b86-409e-9543-1c57e0f4b850-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.125825] env[61868]: DEBUG oslo_concurrency.lockutils [None req-801169df-e652-4a2f-9396-52c991a74fc1 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "4216e9b6-7b86-409e-9543-1c57e0f4b850-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.125994] env[61868]: DEBUG oslo_concurrency.lockutils [None req-801169df-e652-4a2f-9396-52c991a74fc1 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "4216e9b6-7b86-409e-9543-1c57e0f4b850-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.128228] env[61868]: INFO nova.compute.manager [None req-801169df-e652-4a2f-9396-52c991a74fc1 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Terminating instance [ 968.130161] env[61868]: DEBUG nova.compute.manager [None req-801169df-e652-4a2f-9396-52c991a74fc1 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 968.130447] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-801169df-e652-4a2f-9396-52c991a74fc1 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 968.131338] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-511a05cb-7b31-4383-8e38-7b3efc37bd75 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.140096] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-801169df-e652-4a2f-9396-52c991a74fc1 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 968.140327] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b3f4faf3-ff66-49b9-bdec-8c962cc3afd9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.213306] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-801169df-e652-4a2f-9396-52c991a74fc1 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 968.213517] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-801169df-e652-4a2f-9396-52c991a74fc1 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Deleting contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 968.213704] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-801169df-e652-4a2f-9396-52c991a74fc1 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Deleting the datastore file [datastore2] 4216e9b6-7b86-409e-9543-1c57e0f4b850 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 968.213982] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a55249c3-d023-478d-80ab-94ab3979776c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.221606] env[61868]: DEBUG oslo_vmware.api [None req-801169df-e652-4a2f-9396-52c991a74fc1 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 968.221606] env[61868]: value = "task-1315803" [ 968.221606] env[61868]: _type = "Task" [ 968.221606] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.229259] env[61868]: DEBUG oslo_vmware.api [None req-801169df-e652-4a2f-9396-52c991a74fc1 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315803, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.250855] env[61868]: DEBUG nova.compute.manager [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 968.270926] env[61868]: DEBUG nova.virt.hardware [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 968.271174] env[61868]: DEBUG nova.virt.hardware [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 968.271340] env[61868]: DEBUG nova.virt.hardware [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 968.271527] env[61868]: DEBUG nova.virt.hardware [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 968.271677] env[61868]: DEBUG nova.virt.hardware [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 968.271826] env[61868]: DEBUG nova.virt.hardware [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 968.272045] env[61868]: DEBUG nova.virt.hardware [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 968.272262] env[61868]: DEBUG nova.virt.hardware [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 968.272448] env[61868]: DEBUG nova.virt.hardware [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 968.272616] env[61868]: DEBUG nova.virt.hardware [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 968.272788] env[61868]: DEBUG nova.virt.hardware [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 968.273615] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a05d1442-7813-4a87-8ce1-69d47392200c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.281548] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af428fae-dbf1-4506-bb82-0808180b3f59 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.437879] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.208s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.438438] env[61868]: DEBUG nova.compute.manager [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 968.443584] env[61868]: DEBUG oslo_concurrency.lockutils [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.460s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.445060] env[61868]: INFO nova.compute.claims [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 968.461350] env[61868]: DEBUG nova.compute.manager [req-45637716-26b0-4d91-a13d-41bc9dc4ac0e req-47df7ba9-38cf-4ad4-8242-b8dec23baab3 service nova] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Received event network-vif-plugged-4edbd65b-1856-47a2-95c9-e4c8f4d04c71 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 968.461613] env[61868]: DEBUG oslo_concurrency.lockutils [req-45637716-26b0-4d91-a13d-41bc9dc4ac0e req-47df7ba9-38cf-4ad4-8242-b8dec23baab3 service nova] Acquiring lock "9286cc76-1152-4fd4-9279-33d6ee4b9486-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.461821] env[61868]: DEBUG oslo_concurrency.lockutils [req-45637716-26b0-4d91-a13d-41bc9dc4ac0e req-47df7ba9-38cf-4ad4-8242-b8dec23baab3 service nova] Lock "9286cc76-1152-4fd4-9279-33d6ee4b9486-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.461993] env[61868]: DEBUG oslo_concurrency.lockutils [req-45637716-26b0-4d91-a13d-41bc9dc4ac0e req-47df7ba9-38cf-4ad4-8242-b8dec23baab3 service nova] Lock "9286cc76-1152-4fd4-9279-33d6ee4b9486-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.462222] env[61868]: DEBUG nova.compute.manager [req-45637716-26b0-4d91-a13d-41bc9dc4ac0e req-47df7ba9-38cf-4ad4-8242-b8dec23baab3 service nova] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] No waiting events found dispatching network-vif-plugged-4edbd65b-1856-47a2-95c9-e4c8f4d04c71 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 968.462402] env[61868]: WARNING nova.compute.manager [req-45637716-26b0-4d91-a13d-41bc9dc4ac0e req-47df7ba9-38cf-4ad4-8242-b8dec23baab3 service nova] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Received unexpected event network-vif-plugged-4edbd65b-1856-47a2-95c9-e4c8f4d04c71 for instance with vm_state building and task_state spawning. [ 968.571009] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquiring lock "75fdbc11-1ef3-442d-bcf4-55069ba9ab62" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.663447] env[61868]: DEBUG nova.network.neutron [req-611eb4b8-4feb-4851-876e-90189d2eaf05 req-4cc4ea63-4460-4add-b249-288cf6bd9a1f service nova] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Updated VIF entry in instance network info cache for port 21d6cab8-f691-425c-bfa8-9d2f1c294b64. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 968.663783] env[61868]: DEBUG nova.network.neutron [req-611eb4b8-4feb-4851-876e-90189d2eaf05 req-4cc4ea63-4460-4add-b249-288cf6bd9a1f service nova] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Updating instance_info_cache with network_info: [{"id": "21d6cab8-f691-425c-bfa8-9d2f1c294b64", "address": "fa:16:3e:76:a4:70", "network": {"id": "e7457f1a-171a-4c82-a260-1d1a7a95861e", "bridge": null, "label": "tempest-ServersNegativeTestJSON-1032778554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b9b800086764b41a8817d1d8e463a2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap21d6cab8-f6", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.732686] env[61868]: DEBUG oslo_vmware.api [None req-801169df-e652-4a2f-9396-52c991a74fc1 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315803, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.161944} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.733038] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-801169df-e652-4a2f-9396-52c991a74fc1 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 968.733275] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-801169df-e652-4a2f-9396-52c991a74fc1 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Deleted contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 968.733470] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-801169df-e652-4a2f-9396-52c991a74fc1 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 968.733648] env[61868]: INFO nova.compute.manager [None req-801169df-e652-4a2f-9396-52c991a74fc1 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Took 0.60 seconds to destroy the instance on the hypervisor. [ 968.733895] env[61868]: DEBUG oslo.service.loopingcall [None req-801169df-e652-4a2f-9396-52c991a74fc1 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 968.734102] env[61868]: DEBUG nova.compute.manager [-] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 968.734198] env[61868]: DEBUG nova.network.neutron [-] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 968.945467] env[61868]: DEBUG nova.compute.utils [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 968.946714] env[61868]: DEBUG nova.compute.manager [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 968.946928] env[61868]: DEBUG nova.network.neutron [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 968.989761] env[61868]: DEBUG nova.policy [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e02ffdc140274733a081fd53c4acc202', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a60c3c20950e4156b64c3b4c61b9f0f2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 969.045580] env[61868]: DEBUG nova.network.neutron [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Successfully updated port: 4edbd65b-1856-47a2-95c9-e4c8f4d04c71 {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 969.128994] env[61868]: DEBUG nova.objects.instance [None req-97701de6-8f1c-4389-a652-89669d4eb48d tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lazy-loading 'flavor' on Instance uuid e59a182d-97b2-454a-bc40-8afb0839324f {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 969.166474] env[61868]: DEBUG oslo_concurrency.lockutils [req-611eb4b8-4feb-4851-876e-90189d2eaf05 req-4cc4ea63-4460-4add-b249-288cf6bd9a1f service nova] Releasing lock "refresh_cache-75fdbc11-1ef3-442d-bcf4-55069ba9ab62" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.166757] env[61868]: DEBUG nova.compute.manager [req-611eb4b8-4feb-4851-876e-90189d2eaf05 req-4cc4ea63-4460-4add-b249-288cf6bd9a1f service nova] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Received event network-vif-deleted-1f37fd44-6543-44b3-9fce-cb4154f0f0d0 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 969.279290] env[61868]: DEBUG nova.network.neutron [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Successfully created port: 9023a9b8-ad2a-42b6-b2cd-b2cdb86dd9cb {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 969.450665] env[61868]: DEBUG nova.compute.manager [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 969.494632] env[61868]: DEBUG nova.network.neutron [-] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.549588] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "refresh_cache-9286cc76-1152-4fd4-9279-33d6ee4b9486" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.549750] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquired lock "refresh_cache-9286cc76-1152-4fd4-9279-33d6ee4b9486" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.549898] env[61868]: DEBUG nova.network.neutron [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 969.634553] env[61868]: DEBUG oslo_concurrency.lockutils [None req-97701de6-8f1c-4389-a652-89669d4eb48d tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "e59a182d-97b2-454a-bc40-8afb0839324f" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.791s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.677623] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cb30897-3c73-40da-84df-6cf72007acc4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.686235] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f99fb1f-1c88-412f-a3f9-d354bf048c94 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.718251] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41bea29c-026c-4d18-bf95-314423c84832 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.727075] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9860df5-3d09-4940-872d-2b457a051ac3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.740909] env[61868]: DEBUG nova.compute.provider_tree [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 969.998617] env[61868]: INFO nova.compute.manager [-] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Took 1.26 seconds to deallocate network for instance. [ 970.026067] env[61868]: DEBUG nova.compute.manager [req-946f4210-5826-488d-9054-9da6d2309f6b req-b4cba8c4-6923-4c26-addf-d6816ff277a7 service nova] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Received event network-vif-deleted-0f66decc-6560-48cd-8015-0c83c8315e10 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 970.026338] env[61868]: DEBUG nova.compute.manager [req-946f4210-5826-488d-9054-9da6d2309f6b req-b4cba8c4-6923-4c26-addf-d6816ff277a7 service nova] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Received event network-changed-4edbd65b-1856-47a2-95c9-e4c8f4d04c71 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 970.026618] env[61868]: DEBUG nova.compute.manager [req-946f4210-5826-488d-9054-9da6d2309f6b req-b4cba8c4-6923-4c26-addf-d6816ff277a7 service nova] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Refreshing instance network info cache due to event network-changed-4edbd65b-1856-47a2-95c9-e4c8f4d04c71. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 970.026801] env[61868]: DEBUG oslo_concurrency.lockutils [req-946f4210-5826-488d-9054-9da6d2309f6b req-b4cba8c4-6923-4c26-addf-d6816ff277a7 service nova] Acquiring lock "refresh_cache-9286cc76-1152-4fd4-9279-33d6ee4b9486" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.081060] env[61868]: DEBUG nova.network.neutron [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 970.208160] env[61868]: DEBUG nova.network.neutron [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Updating instance_info_cache with network_info: [{"id": "4edbd65b-1856-47a2-95c9-e4c8f4d04c71", "address": "fa:16:3e:18:b8:74", "network": {"id": "77f33b52-e4eb-4b72-85d4-5991a2b5c36d", "bridge": "br-int", "label": "tempest-ServersTestJSON-748635995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b520c3ac58074e8d9b0bfafb817244a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4edbd65b-18", "ovs_interfaceid": "4edbd65b-1856-47a2-95c9-e4c8f4d04c71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.243652] env[61868]: DEBUG nova.scheduler.client.report [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 970.461364] env[61868]: DEBUG nova.compute.manager [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 970.485990] env[61868]: DEBUG nova.virt.hardware [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 970.486283] env[61868]: DEBUG nova.virt.hardware [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 970.486455] env[61868]: DEBUG nova.virt.hardware [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 970.486641] env[61868]: DEBUG nova.virt.hardware [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 970.486792] env[61868]: DEBUG nova.virt.hardware [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 970.486943] env[61868]: DEBUG nova.virt.hardware [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 970.487176] env[61868]: DEBUG nova.virt.hardware [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 970.487346] env[61868]: DEBUG nova.virt.hardware [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 970.487518] env[61868]: DEBUG nova.virt.hardware [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 970.487684] env[61868]: DEBUG nova.virt.hardware [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 970.487859] env[61868]: DEBUG nova.virt.hardware [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 970.488735] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f4ecfdd-4d40-4162-968f-d220e4b25807 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.497649] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b22876d6-b204-4bab-a2e2-8234099ece17 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.511528] env[61868]: DEBUG oslo_concurrency.lockutils [None req-801169df-e652-4a2f-9396-52c991a74fc1 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.572423] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "eeeec752-7e99-471d-9425-c4ccf4f6d0a5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.572661] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "eeeec752-7e99-471d-9425-c4ccf4f6d0a5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.685620] env[61868]: DEBUG nova.compute.manager [req-a61b2244-9a78-4786-a2bf-ff82b464e735 req-506ea058-076b-4930-8dfd-100212b0d8d6 service nova] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Received event network-vif-plugged-9023a9b8-ad2a-42b6-b2cd-b2cdb86dd9cb {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 970.685856] env[61868]: DEBUG oslo_concurrency.lockutils [req-a61b2244-9a78-4786-a2bf-ff82b464e735 req-506ea058-076b-4930-8dfd-100212b0d8d6 service nova] Acquiring lock "5d2b56d5-a4e0-4dba-9158-71036d60a40b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.686077] env[61868]: DEBUG oslo_concurrency.lockutils [req-a61b2244-9a78-4786-a2bf-ff82b464e735 req-506ea058-076b-4930-8dfd-100212b0d8d6 service nova] Lock "5d2b56d5-a4e0-4dba-9158-71036d60a40b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.686255] env[61868]: DEBUG oslo_concurrency.lockutils [req-a61b2244-9a78-4786-a2bf-ff82b464e735 req-506ea058-076b-4930-8dfd-100212b0d8d6 service nova] Lock "5d2b56d5-a4e0-4dba-9158-71036d60a40b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.686461] env[61868]: DEBUG nova.compute.manager [req-a61b2244-9a78-4786-a2bf-ff82b464e735 req-506ea058-076b-4930-8dfd-100212b0d8d6 service nova] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] No waiting events found dispatching network-vif-plugged-9023a9b8-ad2a-42b6-b2cd-b2cdb86dd9cb {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 970.686615] env[61868]: WARNING nova.compute.manager [req-a61b2244-9a78-4786-a2bf-ff82b464e735 req-506ea058-076b-4930-8dfd-100212b0d8d6 service nova] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Received unexpected event network-vif-plugged-9023a9b8-ad2a-42b6-b2cd-b2cdb86dd9cb for instance with vm_state building and task_state spawning. [ 970.710419] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Releasing lock "refresh_cache-9286cc76-1152-4fd4-9279-33d6ee4b9486" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.710713] env[61868]: DEBUG nova.compute.manager [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Instance network_info: |[{"id": "4edbd65b-1856-47a2-95c9-e4c8f4d04c71", "address": "fa:16:3e:18:b8:74", "network": {"id": "77f33b52-e4eb-4b72-85d4-5991a2b5c36d", "bridge": "br-int", "label": "tempest-ServersTestJSON-748635995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b520c3ac58074e8d9b0bfafb817244a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4edbd65b-18", "ovs_interfaceid": "4edbd65b-1856-47a2-95c9-e4c8f4d04c71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 970.710999] env[61868]: DEBUG oslo_concurrency.lockutils [req-946f4210-5826-488d-9054-9da6d2309f6b req-b4cba8c4-6923-4c26-addf-d6816ff277a7 service nova] Acquired lock "refresh_cache-9286cc76-1152-4fd4-9279-33d6ee4b9486" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.711194] env[61868]: DEBUG nova.network.neutron [req-946f4210-5826-488d-9054-9da6d2309f6b req-b4cba8c4-6923-4c26-addf-d6816ff277a7 service nova] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Refreshing network info cache for port 4edbd65b-1856-47a2-95c9-e4c8f4d04c71 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 970.712491] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:18:b8:74', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56136ef6-99d7-4562-9a9f-d66fec951c5c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4edbd65b-1856-47a2-95c9-e4c8f4d04c71', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 970.719843] env[61868]: DEBUG oslo.service.loopingcall [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 970.720940] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 970.721190] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b6b7ceda-1bea-466c-a083-9486e2cb7dae {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.742441] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 970.742441] env[61868]: value = "task-1315804" [ 970.742441] env[61868]: _type = "Task" [ 970.742441] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.747860] env[61868]: DEBUG oslo_concurrency.lockutils [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.305s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.748366] env[61868]: DEBUG nova.compute.manager [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 970.753824] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7de8b1d5-6af1-44ba-9aaf-67e96fc429b9 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.532s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.754046] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7de8b1d5-6af1-44ba-9aaf-67e96fc429b9 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.755961] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.779s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.757380] env[61868]: INFO nova.compute.claims [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 970.760073] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315804, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.787600] env[61868]: INFO nova.scheduler.client.report [None req-7de8b1d5-6af1-44ba-9aaf-67e96fc429b9 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Deleted allocations for instance 842c8606-632a-47d4-9a67-38d8e6ce6b4d [ 970.798254] env[61868]: DEBUG nova.network.neutron [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Successfully updated port: 9023a9b8-ad2a-42b6-b2cd-b2cdb86dd9cb {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 971.075535] env[61868]: DEBUG nova.compute.manager [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 971.253251] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315804, 'name': CreateVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.263956] env[61868]: DEBUG nova.compute.utils [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 971.265302] env[61868]: DEBUG nova.compute.manager [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 971.265466] env[61868]: DEBUG nova.network.neutron [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 971.299164] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7de8b1d5-6af1-44ba-9aaf-67e96fc429b9 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "842c8606-632a-47d4-9a67-38d8e6ce6b4d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.051s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.300394] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "refresh_cache-5d2b56d5-a4e0-4dba-9158-71036d60a40b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.300394] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquired lock "refresh_cache-5d2b56d5-a4e0-4dba-9158-71036d60a40b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.300573] env[61868]: DEBUG nova.network.neutron [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 971.303143] env[61868]: DEBUG nova.policy [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c14b4ff01717495ca97c7f35f91c2995', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd7fc76299baf4a97b57139e5f1caa16e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 971.428886] env[61868]: DEBUG nova.network.neutron [req-946f4210-5826-488d-9054-9da6d2309f6b req-b4cba8c4-6923-4c26-addf-d6816ff277a7 service nova] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Updated VIF entry in instance network info cache for port 4edbd65b-1856-47a2-95c9-e4c8f4d04c71. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 971.429585] env[61868]: DEBUG nova.network.neutron [req-946f4210-5826-488d-9054-9da6d2309f6b req-b4cba8c4-6923-4c26-addf-d6816ff277a7 service nova] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Updating instance_info_cache with network_info: [{"id": "4edbd65b-1856-47a2-95c9-e4c8f4d04c71", "address": "fa:16:3e:18:b8:74", "network": {"id": "77f33b52-e4eb-4b72-85d4-5991a2b5c36d", "bridge": "br-int", "label": "tempest-ServersTestJSON-748635995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b520c3ac58074e8d9b0bfafb817244a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4edbd65b-18", "ovs_interfaceid": "4edbd65b-1856-47a2-95c9-e4c8f4d04c71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.581065] env[61868]: DEBUG nova.network.neutron [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Successfully created port: cd9efdc6-e6f5-4449-b95d-d1fbec53a295 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 971.606356] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.755400] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315804, 'name': CreateVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.768758] env[61868]: DEBUG nova.compute.manager [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 971.850421] env[61868]: DEBUG nova.network.neutron [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 971.922765] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9ebb4ff-3daf-4bec-8dea-ac0fb8dd2754 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.932249] env[61868]: DEBUG oslo_concurrency.lockutils [req-946f4210-5826-488d-9054-9da6d2309f6b req-b4cba8c4-6923-4c26-addf-d6816ff277a7 service nova] Releasing lock "refresh_cache-9286cc76-1152-4fd4-9279-33d6ee4b9486" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.933491] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcadfa00-c3d5-42cb-b3a6-0cb23282fdb2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.972040] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f83e3a5d-cf5a-4f13-bdce-b677b57f220a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.980554] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13ea1551-9796-438d-b47d-9f56eb090029 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.996825] env[61868]: DEBUG nova.compute.provider_tree [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 972.101520] env[61868]: DEBUG nova.network.neutron [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Updating instance_info_cache with network_info: [{"id": "9023a9b8-ad2a-42b6-b2cd-b2cdb86dd9cb", "address": "fa:16:3e:be:bc:fe", "network": {"id": "c1f5bf03-438f-4acc-93fd-159bed7d4069", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2061742146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a60c3c20950e4156b64c3b4c61b9f0f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24b91e3b-50f1-4a16-b929-942e6b31b2d8", "external-id": "nsx-vlan-transportzone-34", "segmentation_id": 34, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9023a9b8-ad", "ovs_interfaceid": "9023a9b8-ad2a-42b6-b2cd-b2cdb86dd9cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.255514] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315804, 'name': CreateVM_Task, 'duration_secs': 1.313593} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.255889] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 972.256457] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.256633] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.257079] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 972.257366] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bdbb5269-52b4-44a0-94bc-be192de0baaf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.262749] env[61868]: DEBUG oslo_vmware.api [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 972.262749] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52285b5f-1427-fbc0-736b-4f081daecea7" [ 972.262749] env[61868]: _type = "Task" [ 972.262749] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.283304] env[61868]: DEBUG oslo_vmware.api [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52285b5f-1427-fbc0-736b-4f081daecea7, 'name': SearchDatastore_Task, 'duration_secs': 0.012164} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.283580] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.283851] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 972.284077] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.284209] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.284394] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 972.284751] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1d72c143-d6a2-4edd-9db3-123e5a4257f0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.293539] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 972.293726] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 972.294621] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3e41447-0600-4607-b619-fd826b42dd79 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.300476] env[61868]: DEBUG oslo_vmware.api [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 972.300476] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52875270-2bd9-9fbc-e1ff-ac77ac1b755a" [ 972.300476] env[61868]: _type = "Task" [ 972.300476] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.309717] env[61868]: DEBUG oslo_vmware.api [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52875270-2bd9-9fbc-e1ff-ac77ac1b755a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.500118] env[61868]: DEBUG nova.scheduler.client.report [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 972.604333] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Releasing lock "refresh_cache-5d2b56d5-a4e0-4dba-9158-71036d60a40b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.604668] env[61868]: DEBUG nova.compute.manager [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Instance network_info: |[{"id": "9023a9b8-ad2a-42b6-b2cd-b2cdb86dd9cb", "address": "fa:16:3e:be:bc:fe", "network": {"id": "c1f5bf03-438f-4acc-93fd-159bed7d4069", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2061742146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a60c3c20950e4156b64c3b4c61b9f0f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24b91e3b-50f1-4a16-b929-942e6b31b2d8", "external-id": "nsx-vlan-transportzone-34", "segmentation_id": 34, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9023a9b8-ad", "ovs_interfaceid": "9023a9b8-ad2a-42b6-b2cd-b2cdb86dd9cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 972.605142] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:be:bc:fe', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24b91e3b-50f1-4a16-b929-942e6b31b2d8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9023a9b8-ad2a-42b6-b2cd-b2cdb86dd9cb', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 972.612698] env[61868]: DEBUG oslo.service.loopingcall [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 972.612918] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 972.613150] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e2bf4089-c664-41fb-a21a-6fadcd30ae11 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.633939] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 972.633939] env[61868]: value = "task-1315805" [ 972.633939] env[61868]: _type = "Task" [ 972.633939] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.641670] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315805, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.719354] env[61868]: DEBUG nova.compute.manager [req-8b042114-11d4-4c8b-bfeb-f3fb84c6bf01 req-e44d738c-c296-4eb5-b4a6-be8d4b0fb833 service nova] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Received event network-changed-9023a9b8-ad2a-42b6-b2cd-b2cdb86dd9cb {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 972.719576] env[61868]: DEBUG nova.compute.manager [req-8b042114-11d4-4c8b-bfeb-f3fb84c6bf01 req-e44d738c-c296-4eb5-b4a6-be8d4b0fb833 service nova] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Refreshing instance network info cache due to event network-changed-9023a9b8-ad2a-42b6-b2cd-b2cdb86dd9cb. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 972.719791] env[61868]: DEBUG oslo_concurrency.lockutils [req-8b042114-11d4-4c8b-bfeb-f3fb84c6bf01 req-e44d738c-c296-4eb5-b4a6-be8d4b0fb833 service nova] Acquiring lock "refresh_cache-5d2b56d5-a4e0-4dba-9158-71036d60a40b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.719941] env[61868]: DEBUG oslo_concurrency.lockutils [req-8b042114-11d4-4c8b-bfeb-f3fb84c6bf01 req-e44d738c-c296-4eb5-b4a6-be8d4b0fb833 service nova] Acquired lock "refresh_cache-5d2b56d5-a4e0-4dba-9158-71036d60a40b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.720120] env[61868]: DEBUG nova.network.neutron [req-8b042114-11d4-4c8b-bfeb-f3fb84c6bf01 req-e44d738c-c296-4eb5-b4a6-be8d4b0fb833 service nova] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Refreshing network info cache for port 9023a9b8-ad2a-42b6-b2cd-b2cdb86dd9cb {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 972.777849] env[61868]: DEBUG nova.compute.manager [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 972.806373] env[61868]: DEBUG nova.virt.hardware [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 972.806664] env[61868]: DEBUG nova.virt.hardware [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 972.806829] env[61868]: DEBUG nova.virt.hardware [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 972.807033] env[61868]: DEBUG nova.virt.hardware [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 972.807193] env[61868]: DEBUG nova.virt.hardware [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 972.807346] env[61868]: DEBUG nova.virt.hardware [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 972.807555] env[61868]: DEBUG nova.virt.hardware [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 972.807717] env[61868]: DEBUG nova.virt.hardware [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 972.807885] env[61868]: DEBUG nova.virt.hardware [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 972.808061] env[61868]: DEBUG nova.virt.hardware [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 972.808308] env[61868]: DEBUG nova.virt.hardware [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 972.809060] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eaf3de7-91f0-4b87-814e-23b23bb5405e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.820358] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09ffb348-1a49-4007-b722-59ff71c57861 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.824041] env[61868]: DEBUG oslo_vmware.api [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52875270-2bd9-9fbc-e1ff-ac77ac1b755a, 'name': SearchDatastore_Task, 'duration_secs': 0.00934} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.825144] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14275373-aab2-4319-894d-e7e9e19a74f8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.836489] env[61868]: DEBUG oslo_vmware.api [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 972.836489] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5212b41f-3086-87e9-dd94-7bb396b247d4" [ 972.836489] env[61868]: _type = "Task" [ 972.836489] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.844074] env[61868]: DEBUG oslo_vmware.api [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5212b41f-3086-87e9-dd94-7bb396b247d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.005984] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.250s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.006611] env[61868]: DEBUG nova.compute.manager [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 973.009480] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.883s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.009707] env[61868]: DEBUG nova.objects.instance [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lazy-loading 'resources' on Instance uuid 75fdbc11-1ef3-442d-bcf4-55069ba9ab62 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 973.144521] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315805, 'name': CreateVM_Task, 'duration_secs': 0.307231} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.144716] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 973.145553] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.145773] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.146336] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 973.146445] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9915c495-6230-445c-9922-015949fe662b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.151620] env[61868]: DEBUG oslo_vmware.api [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 973.151620] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52bdef6e-69c1-ddfb-3bbb-19a8464a02d5" [ 973.151620] env[61868]: _type = "Task" [ 973.151620] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.159963] env[61868]: DEBUG oslo_vmware.api [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52bdef6e-69c1-ddfb-3bbb-19a8464a02d5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.231017] env[61868]: DEBUG nova.compute.manager [req-e658819e-3ba1-4082-a24d-fba82199e01e req-d478a85d-8445-4e08-a417-07f21af35dc2 service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Received event network-vif-plugged-cd9efdc6-e6f5-4449-b95d-d1fbec53a295 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 973.231017] env[61868]: DEBUG oslo_concurrency.lockutils [req-e658819e-3ba1-4082-a24d-fba82199e01e req-d478a85d-8445-4e08-a417-07f21af35dc2 service nova] Acquiring lock "a96cab5a-c053-4e45-96f6-2aba0a819110-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.231017] env[61868]: DEBUG oslo_concurrency.lockutils [req-e658819e-3ba1-4082-a24d-fba82199e01e req-d478a85d-8445-4e08-a417-07f21af35dc2 service nova] Lock "a96cab5a-c053-4e45-96f6-2aba0a819110-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.231017] env[61868]: DEBUG oslo_concurrency.lockutils [req-e658819e-3ba1-4082-a24d-fba82199e01e req-d478a85d-8445-4e08-a417-07f21af35dc2 service nova] Lock "a96cab5a-c053-4e45-96f6-2aba0a819110-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.231017] env[61868]: DEBUG nova.compute.manager [req-e658819e-3ba1-4082-a24d-fba82199e01e req-d478a85d-8445-4e08-a417-07f21af35dc2 service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] No waiting events found dispatching network-vif-plugged-cd9efdc6-e6f5-4449-b95d-d1fbec53a295 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 973.231017] env[61868]: WARNING nova.compute.manager [req-e658819e-3ba1-4082-a24d-fba82199e01e req-d478a85d-8445-4e08-a417-07f21af35dc2 service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Received unexpected event network-vif-plugged-cd9efdc6-e6f5-4449-b95d-d1fbec53a295 for instance with vm_state building and task_state spawning. [ 973.277097] env[61868]: DEBUG nova.network.neutron [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Successfully updated port: cd9efdc6-e6f5-4449-b95d-d1fbec53a295 {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 973.349724] env[61868]: DEBUG oslo_vmware.api [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5212b41f-3086-87e9-dd94-7bb396b247d4, 'name': SearchDatastore_Task, 'duration_secs': 0.013186} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.352142] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.352473] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 9286cc76-1152-4fd4-9279-33d6ee4b9486/9286cc76-1152-4fd4-9279-33d6ee4b9486.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 973.352759] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b9e7e11d-2729-455b-813c-4fa8ad006ecc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.362234] env[61868]: DEBUG oslo_vmware.api [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 973.362234] env[61868]: value = "task-1315806" [ 973.362234] env[61868]: _type = "Task" [ 973.362234] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.370827] env[61868]: DEBUG oslo_vmware.api [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315806, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.512661] env[61868]: DEBUG nova.compute.utils [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 973.514663] env[61868]: DEBUG nova.objects.instance [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lazy-loading 'numa_topology' on Instance uuid 75fdbc11-1ef3-442d-bcf4-55069ba9ab62 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 973.515904] env[61868]: DEBUG nova.compute.manager [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 973.516091] env[61868]: DEBUG nova.network.neutron [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 973.531859] env[61868]: DEBUG nova.network.neutron [req-8b042114-11d4-4c8b-bfeb-f3fb84c6bf01 req-e44d738c-c296-4eb5-b4a6-be8d4b0fb833 service nova] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Updated VIF entry in instance network info cache for port 9023a9b8-ad2a-42b6-b2cd-b2cdb86dd9cb. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 973.532358] env[61868]: DEBUG nova.network.neutron [req-8b042114-11d4-4c8b-bfeb-f3fb84c6bf01 req-e44d738c-c296-4eb5-b4a6-be8d4b0fb833 service nova] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Updating instance_info_cache with network_info: [{"id": "9023a9b8-ad2a-42b6-b2cd-b2cdb86dd9cb", "address": "fa:16:3e:be:bc:fe", "network": {"id": "c1f5bf03-438f-4acc-93fd-159bed7d4069", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2061742146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a60c3c20950e4156b64c3b4c61b9f0f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24b91e3b-50f1-4a16-b929-942e6b31b2d8", "external-id": "nsx-vlan-transportzone-34", "segmentation_id": 34, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9023a9b8-ad", "ovs_interfaceid": "9023a9b8-ad2a-42b6-b2cd-b2cdb86dd9cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.561630] env[61868]: DEBUG nova.policy [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0b615dd3d1064706bea31b52fbfa9a61', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5658848e1b0e42929c04a4a8de40a291', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 973.663115] env[61868]: DEBUG oslo_vmware.api [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52bdef6e-69c1-ddfb-3bbb-19a8464a02d5, 'name': SearchDatastore_Task, 'duration_secs': 0.009908} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.663534] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.663785] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 973.664033] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.664189] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.664392] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 973.664731] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-429d0137-b5f0-41a1-8c80-44c47f66d259 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.683928] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 973.684207] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 973.685503] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eee98449-b2a7-499b-9120-7db1d01ef25f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.692223] env[61868]: DEBUG oslo_vmware.api [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 973.692223] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]524c9ec9-406f-dfef-d71e-db9ab19cb534" [ 973.692223] env[61868]: _type = "Task" [ 973.692223] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.700937] env[61868]: DEBUG oslo_vmware.api [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]524c9ec9-406f-dfef-d71e-db9ab19cb534, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.779595] env[61868]: DEBUG oslo_concurrency.lockutils [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "refresh_cache-a96cab5a-c053-4e45-96f6-2aba0a819110" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.779794] env[61868]: DEBUG oslo_concurrency.lockutils [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquired lock "refresh_cache-a96cab5a-c053-4e45-96f6-2aba0a819110" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.779949] env[61868]: DEBUG nova.network.neutron [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 973.834593] env[61868]: DEBUG nova.network.neutron [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Successfully created port: 5736bca4-8f2f-485a-bc60-41a19416448c {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 973.872116] env[61868]: DEBUG oslo_vmware.api [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315806, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.450214} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.872400] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 9286cc76-1152-4fd4-9279-33d6ee4b9486/9286cc76-1152-4fd4-9279-33d6ee4b9486.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 973.872618] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 973.872864] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8ba68cd9-ab1f-4715-9a0b-4d33a87a7374 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.881085] env[61868]: DEBUG oslo_vmware.api [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 973.881085] env[61868]: value = "task-1315807" [ 973.881085] env[61868]: _type = "Task" [ 973.881085] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.890239] env[61868]: DEBUG oslo_vmware.api [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315807, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.016761] env[61868]: DEBUG nova.compute.manager [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 974.019533] env[61868]: DEBUG nova.objects.base [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Object Instance<75fdbc11-1ef3-442d-bcf4-55069ba9ab62> lazy-loaded attributes: resources,numa_topology {{(pid=61868) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 974.034825] env[61868]: DEBUG oslo_concurrency.lockutils [req-8b042114-11d4-4c8b-bfeb-f3fb84c6bf01 req-e44d738c-c296-4eb5-b4a6-be8d4b0fb833 service nova] Releasing lock "refresh_cache-5d2b56d5-a4e0-4dba-9158-71036d60a40b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.144187] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97fdff24-7cf4-467a-81ac-ceb87c419529 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.152483] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-396bf87b-ff1e-49e4-b7ad-fd3d1db473e8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.184478] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc8100ac-4aae-4c5a-9d34-95d8b0cde4c9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.192718] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cb0318c-c6d7-412a-9dac-5c9bbcdef646 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.205266] env[61868]: DEBUG oslo_vmware.api [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]524c9ec9-406f-dfef-d71e-db9ab19cb534, 'name': SearchDatastore_Task, 'duration_secs': 0.055503} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.213668] env[61868]: DEBUG nova.compute.provider_tree [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 974.214995] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce907f81-ba4c-4c7e-a8da-ad0f9c262269 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.220780] env[61868]: DEBUG oslo_vmware.api [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 974.220780] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52fb2989-dc21-0212-e446-b1b8e84b61ab" [ 974.220780] env[61868]: _type = "Task" [ 974.220780] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.230658] env[61868]: DEBUG oslo_vmware.api [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52fb2989-dc21-0212-e446-b1b8e84b61ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.325968] env[61868]: DEBUG nova.network.neutron [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 974.396236] env[61868]: DEBUG oslo_vmware.api [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315807, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072637} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.399249] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 974.400116] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10404499-06e2-4f44-9fe9-559919e8f9a0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.425812] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] 9286cc76-1152-4fd4-9279-33d6ee4b9486/9286cc76-1152-4fd4-9279-33d6ee4b9486.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 974.426177] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-25d5fc01-8094-47c3-9984-96a23ab5c518 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.448243] env[61868]: DEBUG oslo_vmware.api [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 974.448243] env[61868]: value = "task-1315808" [ 974.448243] env[61868]: _type = "Task" [ 974.448243] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.455887] env[61868]: DEBUG oslo_vmware.api [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315808, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.553235] env[61868]: DEBUG nova.network.neutron [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Updating instance_info_cache with network_info: [{"id": "cd9efdc6-e6f5-4449-b95d-d1fbec53a295", "address": "fa:16:3e:bf:13:19", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd9efdc6-e6", "ovs_interfaceid": "cd9efdc6-e6f5-4449-b95d-d1fbec53a295", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.719063] env[61868]: DEBUG nova.scheduler.client.report [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 974.735020] env[61868]: DEBUG oslo_vmware.api [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52fb2989-dc21-0212-e446-b1b8e84b61ab, 'name': SearchDatastore_Task, 'duration_secs': 0.010156} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.735872] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.736157] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 5d2b56d5-a4e0-4dba-9158-71036d60a40b/5d2b56d5-a4e0-4dba-9158-71036d60a40b.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 974.736427] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2df42d24-a5bf-4c9f-80f8-738267bee610 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.743693] env[61868]: DEBUG oslo_vmware.api [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 974.743693] env[61868]: value = "task-1315809" [ 974.743693] env[61868]: _type = "Task" [ 974.743693] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.751613] env[61868]: DEBUG oslo_vmware.api [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315809, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.958693] env[61868]: DEBUG oslo_vmware.api [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315808, 'name': ReconfigVM_Task, 'duration_secs': 0.273264} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.959143] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Reconfigured VM instance instance-00000059 to attach disk [datastore1] 9286cc76-1152-4fd4-9279-33d6ee4b9486/9286cc76-1152-4fd4-9279-33d6ee4b9486.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 974.959811] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2240434f-7aa4-4620-ac3f-c0c0b45f592b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.968488] env[61868]: DEBUG oslo_vmware.api [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 974.968488] env[61868]: value = "task-1315810" [ 974.968488] env[61868]: _type = "Task" [ 974.968488] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.977608] env[61868]: DEBUG oslo_vmware.api [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315810, 'name': Rename_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.030012] env[61868]: DEBUG nova.compute.manager [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 975.056651] env[61868]: DEBUG nova.virt.hardware [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 975.056907] env[61868]: DEBUG nova.virt.hardware [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 975.057120] env[61868]: DEBUG nova.virt.hardware [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 975.057337] env[61868]: DEBUG nova.virt.hardware [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 975.057492] env[61868]: DEBUG nova.virt.hardware [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 975.057643] env[61868]: DEBUG nova.virt.hardware [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 975.057887] env[61868]: DEBUG nova.virt.hardware [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 975.058102] env[61868]: DEBUG nova.virt.hardware [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 975.058294] env[61868]: DEBUG nova.virt.hardware [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 975.058467] env[61868]: DEBUG nova.virt.hardware [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 975.058647] env[61868]: DEBUG nova.virt.hardware [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 975.059175] env[61868]: DEBUG oslo_concurrency.lockutils [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Releasing lock "refresh_cache-a96cab5a-c053-4e45-96f6-2aba0a819110" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.059496] env[61868]: DEBUG nova.compute.manager [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Instance network_info: |[{"id": "cd9efdc6-e6f5-4449-b95d-d1fbec53a295", "address": "fa:16:3e:bf:13:19", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd9efdc6-e6", "ovs_interfaceid": "cd9efdc6-e6f5-4449-b95d-d1fbec53a295", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 975.060416] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7981e3ae-e217-426c-b0e8-c9fcc9c62a07 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.063494] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bf:13:19', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8b29df12-5674-476d-a9e5-5e20f704d224', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cd9efdc6-e6f5-4449-b95d-d1fbec53a295', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 975.070769] env[61868]: DEBUG oslo.service.loopingcall [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 975.071022] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 975.071761] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-074983b5-1e18-4c39-a896-a20e68a4fd99 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.090796] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e26be353-d221-4260-934e-7ada82a0ff50 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.097562] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 975.097562] env[61868]: value = "task-1315811" [ 975.097562] env[61868]: _type = "Task" [ 975.097562] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.119393] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315811, 'name': CreateVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.228266] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.218s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.232233] env[61868]: DEBUG oslo_concurrency.lockutils [None req-39bcf7a5-825e-47cc-9e88-e67d23c1c42e tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.686s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.232558] env[61868]: DEBUG nova.objects.instance [None req-39bcf7a5-825e-47cc-9e88-e67d23c1c42e tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lazy-loading 'resources' on Instance uuid 867384c6-175b-4848-b9ac-917470742ba2 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 975.256390] env[61868]: DEBUG oslo_vmware.api [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315809, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.486268} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.256724] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 5d2b56d5-a4e0-4dba-9158-71036d60a40b/5d2b56d5-a4e0-4dba-9158-71036d60a40b.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 975.256952] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 975.257350] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bc33463b-72ce-453e-9a9e-f8ba1edfa781 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.273596] env[61868]: DEBUG nova.compute.manager [req-34c5af64-8420-48ff-8685-7a6c60ea8d10 req-932e58a6-68e6-428e-9050-adc29382d1c8 service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Received event network-changed-cd9efdc6-e6f5-4449-b95d-d1fbec53a295 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 975.273802] env[61868]: DEBUG nova.compute.manager [req-34c5af64-8420-48ff-8685-7a6c60ea8d10 req-932e58a6-68e6-428e-9050-adc29382d1c8 service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Refreshing instance network info cache due to event network-changed-cd9efdc6-e6f5-4449-b95d-d1fbec53a295. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 975.274091] env[61868]: DEBUG oslo_concurrency.lockutils [req-34c5af64-8420-48ff-8685-7a6c60ea8d10 req-932e58a6-68e6-428e-9050-adc29382d1c8 service nova] Acquiring lock "refresh_cache-a96cab5a-c053-4e45-96f6-2aba0a819110" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.274278] env[61868]: DEBUG oslo_concurrency.lockutils [req-34c5af64-8420-48ff-8685-7a6c60ea8d10 req-932e58a6-68e6-428e-9050-adc29382d1c8 service nova] Acquired lock "refresh_cache-a96cab5a-c053-4e45-96f6-2aba0a819110" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.274454] env[61868]: DEBUG nova.network.neutron [req-34c5af64-8420-48ff-8685-7a6c60ea8d10 req-932e58a6-68e6-428e-9050-adc29382d1c8 service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Refreshing network info cache for port cd9efdc6-e6f5-4449-b95d-d1fbec53a295 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 975.279603] env[61868]: DEBUG oslo_vmware.api [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 975.279603] env[61868]: value = "task-1315812" [ 975.279603] env[61868]: _type = "Task" [ 975.279603] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.289095] env[61868]: DEBUG oslo_vmware.api [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315812, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.355247] env[61868]: DEBUG nova.network.neutron [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Successfully updated port: 5736bca4-8f2f-485a-bc60-41a19416448c {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 975.479743] env[61868]: DEBUG oslo_vmware.api [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315810, 'name': Rename_Task, 'duration_secs': 0.270824} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.480035] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 975.480314] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2be87fb7-131f-4128-b631-405b49eac6d1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.487061] env[61868]: DEBUG oslo_vmware.api [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 975.487061] env[61868]: value = "task-1315813" [ 975.487061] env[61868]: _type = "Task" [ 975.487061] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.494715] env[61868]: DEBUG oslo_vmware.api [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315813, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.608463] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315811, 'name': CreateVM_Task, 'duration_secs': 0.402634} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.608655] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 975.609370] env[61868]: DEBUG oslo_concurrency.lockutils [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.609545] env[61868]: DEBUG oslo_concurrency.lockutils [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.609871] env[61868]: DEBUG oslo_concurrency.lockutils [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 975.610178] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dbda9804-8843-47c5-9813-acc31de76a02 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.616044] env[61868]: DEBUG oslo_vmware.api [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for the task: (returnval){ [ 975.616044] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]527330da-b3a0-c1fd-f969-d439717d960e" [ 975.616044] env[61868]: _type = "Task" [ 975.616044] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.623399] env[61868]: DEBUG oslo_vmware.api [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]527330da-b3a0-c1fd-f969-d439717d960e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.742316] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ca67f30a-4542-47f7-b8ad-1b6a7c66a8e4 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lock "75fdbc11-1ef3-442d-bcf4-55069ba9ab62" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 26.820s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.743208] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lock "75fdbc11-1ef3-442d-bcf4-55069ba9ab62" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 7.172s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.743810] env[61868]: INFO nova.compute.manager [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Unshelving [ 975.788944] env[61868]: DEBUG oslo_vmware.api [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315812, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084123} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.791320] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 975.792297] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e434eae-654f-4ad1-bd14-71ff91df35c2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.815263] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] 5d2b56d5-a4e0-4dba-9158-71036d60a40b/5d2b56d5-a4e0-4dba-9158-71036d60a40b.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 975.819656] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e857bf18-3aab-46d2-8c3a-f6a67d85d9ef {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.841734] env[61868]: DEBUG oslo_vmware.api [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 975.841734] env[61868]: value = "task-1315814" [ 975.841734] env[61868]: _type = "Task" [ 975.841734] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.851170] env[61868]: DEBUG oslo_vmware.api [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315814, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.858143] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "refresh_cache-cbeeb7b1-166b-49f7-88c5-6ed05c9dc759" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.858370] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquired lock "refresh_cache-cbeeb7b1-166b-49f7-88c5-6ed05c9dc759" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.858545] env[61868]: DEBUG nova.network.neutron [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 975.912968] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8d7c752-a182-43c0-bbad-fcf38caa8061 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.922181] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7797a131-7f16-4120-b971-0395ec4c901b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.960167] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d77c5137-e2db-460e-9569-4e96b0bc9d17 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.968231] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4637dc1d-2a5f-4fa9-9495-1a6223129b15 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.981986] env[61868]: DEBUG nova.compute.provider_tree [None req-39bcf7a5-825e-47cc-9e88-e67d23c1c42e tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 975.997305] env[61868]: DEBUG oslo_vmware.api [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315813, 'name': PowerOnVM_Task, 'duration_secs': 0.448334} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.999539] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 975.999747] env[61868]: INFO nova.compute.manager [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Took 7.75 seconds to spawn the instance on the hypervisor. [ 975.999927] env[61868]: DEBUG nova.compute.manager [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 976.000766] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53f1be44-a8b0-4f40-ba1e-21967a29a44b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.036070] env[61868]: DEBUG nova.network.neutron [req-34c5af64-8420-48ff-8685-7a6c60ea8d10 req-932e58a6-68e6-428e-9050-adc29382d1c8 service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Updated VIF entry in instance network info cache for port cd9efdc6-e6f5-4449-b95d-d1fbec53a295. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 976.036444] env[61868]: DEBUG nova.network.neutron [req-34c5af64-8420-48ff-8685-7a6c60ea8d10 req-932e58a6-68e6-428e-9050-adc29382d1c8 service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Updating instance_info_cache with network_info: [{"id": "cd9efdc6-e6f5-4449-b95d-d1fbec53a295", "address": "fa:16:3e:bf:13:19", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd9efdc6-e6", "ovs_interfaceid": "cd9efdc6-e6f5-4449-b95d-d1fbec53a295", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.125563] env[61868]: DEBUG oslo_vmware.api [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]527330da-b3a0-c1fd-f969-d439717d960e, 'name': SearchDatastore_Task, 'duration_secs': 0.014141} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.125878] env[61868]: DEBUG oslo_concurrency.lockutils [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.126147] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 976.126427] env[61868]: DEBUG oslo_concurrency.lockutils [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.126540] env[61868]: DEBUG oslo_concurrency.lockutils [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.126722] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 976.126978] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-53badb9f-63da-4007-8141-8e91c2df2666 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.140546] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 976.140725] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 976.141805] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a12d8e2-5375-46bb-b93f-7b8ba5aa8b5d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.146627] env[61868]: DEBUG oslo_vmware.api [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for the task: (returnval){ [ 976.146627] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]528eddf4-cd86-dc24-1c3a-87b057677e3e" [ 976.146627] env[61868]: _type = "Task" [ 976.146627] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.154265] env[61868]: DEBUG oslo_vmware.api [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]528eddf4-cd86-dc24-1c3a-87b057677e3e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.356384] env[61868]: DEBUG oslo_vmware.api [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315814, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.410571] env[61868]: DEBUG nova.network.neutron [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 976.486015] env[61868]: DEBUG nova.scheduler.client.report [None req-39bcf7a5-825e-47cc-9e88-e67d23c1c42e tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 976.519444] env[61868]: INFO nova.compute.manager [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Took 12.52 seconds to build instance. [ 976.539064] env[61868]: DEBUG oslo_concurrency.lockutils [req-34c5af64-8420-48ff-8685-7a6c60ea8d10 req-932e58a6-68e6-428e-9050-adc29382d1c8 service nova] Releasing lock "refresh_cache-a96cab5a-c053-4e45-96f6-2aba0a819110" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.539328] env[61868]: DEBUG nova.compute.manager [req-34c5af64-8420-48ff-8685-7a6c60ea8d10 req-932e58a6-68e6-428e-9050-adc29382d1c8 service nova] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Received event network-vif-plugged-5736bca4-8f2f-485a-bc60-41a19416448c {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 976.539523] env[61868]: DEBUG oslo_concurrency.lockutils [req-34c5af64-8420-48ff-8685-7a6c60ea8d10 req-932e58a6-68e6-428e-9050-adc29382d1c8 service nova] Acquiring lock "cbeeb7b1-166b-49f7-88c5-6ed05c9dc759-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.539762] env[61868]: DEBUG oslo_concurrency.lockutils [req-34c5af64-8420-48ff-8685-7a6c60ea8d10 req-932e58a6-68e6-428e-9050-adc29382d1c8 service nova] Lock "cbeeb7b1-166b-49f7-88c5-6ed05c9dc759-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.539935] env[61868]: DEBUG oslo_concurrency.lockutils [req-34c5af64-8420-48ff-8685-7a6c60ea8d10 req-932e58a6-68e6-428e-9050-adc29382d1c8 service nova] Lock "cbeeb7b1-166b-49f7-88c5-6ed05c9dc759-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.540124] env[61868]: DEBUG nova.compute.manager [req-34c5af64-8420-48ff-8685-7a6c60ea8d10 req-932e58a6-68e6-428e-9050-adc29382d1c8 service nova] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] No waiting events found dispatching network-vif-plugged-5736bca4-8f2f-485a-bc60-41a19416448c {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 976.540302] env[61868]: WARNING nova.compute.manager [req-34c5af64-8420-48ff-8685-7a6c60ea8d10 req-932e58a6-68e6-428e-9050-adc29382d1c8 service nova] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Received unexpected event network-vif-plugged-5736bca4-8f2f-485a-bc60-41a19416448c for instance with vm_state building and task_state spawning. [ 976.557699] env[61868]: DEBUG nova.network.neutron [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Updating instance_info_cache with network_info: [{"id": "5736bca4-8f2f-485a-bc60-41a19416448c", "address": "fa:16:3e:f2:7e:d2", "network": {"id": "f2e3f911-d826-43ad-911d-9f6985db7ee5", "bridge": "br-int", "label": "tempest-ImagesTestJSON-485234144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5658848e1b0e42929c04a4a8de40a291", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5736bca4-8f", "ovs_interfaceid": "5736bca4-8f2f-485a-bc60-41a19416448c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.657156] env[61868]: DEBUG oslo_vmware.api [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]528eddf4-cd86-dc24-1c3a-87b057677e3e, 'name': SearchDatastore_Task, 'duration_secs': 0.050904} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.657910] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b9742ef9-a471-4486-8a06-15110307ca30 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.663970] env[61868]: DEBUG oslo_vmware.api [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for the task: (returnval){ [ 976.663970] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52215acd-942f-fae6-a92c-f871780e4beb" [ 976.663970] env[61868]: _type = "Task" [ 976.663970] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.673136] env[61868]: DEBUG oslo_vmware.api [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52215acd-942f-fae6-a92c-f871780e4beb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.766475] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.851699] env[61868]: DEBUG oslo_vmware.api [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315814, 'name': ReconfigVM_Task, 'duration_secs': 0.663466} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.851982] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Reconfigured VM instance instance-00000058 to attach disk [datastore1] 5d2b56d5-a4e0-4dba-9158-71036d60a40b/5d2b56d5-a4e0-4dba-9158-71036d60a40b.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 976.852677] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e8ef7638-ba07-4d9e-af69-85aa7be21bb2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.859066] env[61868]: DEBUG oslo_vmware.api [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 976.859066] env[61868]: value = "task-1315815" [ 976.859066] env[61868]: _type = "Task" [ 976.859066] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.867497] env[61868]: DEBUG oslo_vmware.api [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315815, 'name': Rename_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.990858] env[61868]: DEBUG oslo_concurrency.lockutils [None req-39bcf7a5-825e-47cc-9e88-e67d23c1c42e tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.759s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.994062] env[61868]: DEBUG oslo_concurrency.lockutils [None req-801169df-e652-4a2f-9396-52c991a74fc1 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.482s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.994062] env[61868]: DEBUG nova.objects.instance [None req-801169df-e652-4a2f-9396-52c991a74fc1 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lazy-loading 'resources' on Instance uuid 4216e9b6-7b86-409e-9543-1c57e0f4b850 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 977.011398] env[61868]: INFO nova.scheduler.client.report [None req-39bcf7a5-825e-47cc-9e88-e67d23c1c42e tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Deleted allocations for instance 867384c6-175b-4848-b9ac-917470742ba2 [ 977.021707] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6b53fa1f-c971-4bd9-9b8c-38e5546b7b91 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "9286cc76-1152-4fd4-9279-33d6ee4b9486" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.034s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.060679] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Releasing lock "refresh_cache-cbeeb7b1-166b-49f7-88c5-6ed05c9dc759" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.061109] env[61868]: DEBUG nova.compute.manager [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Instance network_info: |[{"id": "5736bca4-8f2f-485a-bc60-41a19416448c", "address": "fa:16:3e:f2:7e:d2", "network": {"id": "f2e3f911-d826-43ad-911d-9f6985db7ee5", "bridge": "br-int", "label": "tempest-ImagesTestJSON-485234144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5658848e1b0e42929c04a4a8de40a291", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5736bca4-8f", "ovs_interfaceid": "5736bca4-8f2f-485a-bc60-41a19416448c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 977.061600] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f2:7e:d2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f1b507ed-cd2d-4c09-9d96-c47bde6a7774', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5736bca4-8f2f-485a-bc60-41a19416448c', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 977.069437] env[61868]: DEBUG oslo.service.loopingcall [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 977.070444] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 977.070729] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-711edfa9-4734-4c91-aeb4-430a9f7d1cfd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.093053] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 977.093053] env[61868]: value = "task-1315816" [ 977.093053] env[61868]: _type = "Task" [ 977.093053] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.101652] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315816, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.175741] env[61868]: DEBUG oslo_vmware.api [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52215acd-942f-fae6-a92c-f871780e4beb, 'name': SearchDatastore_Task, 'duration_secs': 0.008986} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.176026] env[61868]: DEBUG oslo_concurrency.lockutils [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.176295] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] a96cab5a-c053-4e45-96f6-2aba0a819110/a96cab5a-c053-4e45-96f6-2aba0a819110.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 977.176565] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c69b6875-3b86-4292-9d5d-ab45d8a7ffa7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.184946] env[61868]: DEBUG oslo_vmware.api [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for the task: (returnval){ [ 977.184946] env[61868]: value = "task-1315817" [ 977.184946] env[61868]: _type = "Task" [ 977.184946] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.194735] env[61868]: DEBUG oslo_vmware.api [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315817, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.301952] env[61868]: DEBUG nova.compute.manager [req-8ae1736e-191f-4ee8-8820-0ca0efa51ec1 req-b2128c5a-0354-437d-bd7d-aa26f3ecfe61 service nova] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Received event network-changed-5736bca4-8f2f-485a-bc60-41a19416448c {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 977.302217] env[61868]: DEBUG nova.compute.manager [req-8ae1736e-191f-4ee8-8820-0ca0efa51ec1 req-b2128c5a-0354-437d-bd7d-aa26f3ecfe61 service nova] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Refreshing instance network info cache due to event network-changed-5736bca4-8f2f-485a-bc60-41a19416448c. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 977.302419] env[61868]: DEBUG oslo_concurrency.lockutils [req-8ae1736e-191f-4ee8-8820-0ca0efa51ec1 req-b2128c5a-0354-437d-bd7d-aa26f3ecfe61 service nova] Acquiring lock "refresh_cache-cbeeb7b1-166b-49f7-88c5-6ed05c9dc759" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.302557] env[61868]: DEBUG oslo_concurrency.lockutils [req-8ae1736e-191f-4ee8-8820-0ca0efa51ec1 req-b2128c5a-0354-437d-bd7d-aa26f3ecfe61 service nova] Acquired lock "refresh_cache-cbeeb7b1-166b-49f7-88c5-6ed05c9dc759" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.302809] env[61868]: DEBUG nova.network.neutron [req-8ae1736e-191f-4ee8-8820-0ca0efa51ec1 req-b2128c5a-0354-437d-bd7d-aa26f3ecfe61 service nova] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Refreshing network info cache for port 5736bca4-8f2f-485a-bc60-41a19416448c {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 977.370466] env[61868]: DEBUG oslo_vmware.api [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315815, 'name': Rename_Task, 'duration_secs': 0.143747} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.370827] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 977.371227] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-12929f92-a2f4-4bb8-825d-1ef9d6438248 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.378386] env[61868]: DEBUG oslo_vmware.api [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 977.378386] env[61868]: value = "task-1315818" [ 977.378386] env[61868]: _type = "Task" [ 977.378386] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.390891] env[61868]: DEBUG oslo_vmware.api [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315818, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.521676] env[61868]: DEBUG oslo_concurrency.lockutils [None req-39bcf7a5-825e-47cc-9e88-e67d23c1c42e tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "867384c6-175b-4848-b9ac-917470742ba2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.047s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.610075] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315816, 'name': CreateVM_Task, 'duration_secs': 0.385259} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.610158] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 977.610961] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.611154] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.611568] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 977.614831] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef901039-1dee-4533-9fa6-e60c97cddac2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.621962] env[61868]: DEBUG oslo_vmware.api [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 977.621962] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e816c2-18b1-39f7-2dac-c8444e32b947" [ 977.621962] env[61868]: _type = "Task" [ 977.621962] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.633575] env[61868]: DEBUG oslo_vmware.api [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e816c2-18b1-39f7-2dac-c8444e32b947, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.642966] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fa55b02-4a2c-4282-90c9-80f3caa23144 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.652367] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ca7c4b2-adaf-42cf-a3dd-fe0b1f5ce7c8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.693251] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5afd0bae-81aa-4b93-851c-464432515c19 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.704853] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0ca54aa-8779-403a-a786-d0ffa346d4c8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.708662] env[61868]: DEBUG oslo_vmware.api [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315817, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.504502} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.708917] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] a96cab5a-c053-4e45-96f6-2aba0a819110/a96cab5a-c053-4e45-96f6-2aba0a819110.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 977.709138] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 977.709717] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-586082bf-6c19-43b6-87bb-f5fe9aa087d1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.720082] env[61868]: DEBUG nova.compute.provider_tree [None req-801169df-e652-4a2f-9396-52c991a74fc1 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 977.727330] env[61868]: DEBUG oslo_vmware.api [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for the task: (returnval){ [ 977.727330] env[61868]: value = "task-1315819" [ 977.727330] env[61868]: _type = "Task" [ 977.727330] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.727330] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "56703761-3421-44fd-aa63-9db36c86e69c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.727632] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "56703761-3421-44fd-aa63-9db36c86e69c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.740164] env[61868]: DEBUG oslo_vmware.api [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315819, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.888893] env[61868]: DEBUG oslo_vmware.api [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315818, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.007131] env[61868]: DEBUG nova.network.neutron [req-8ae1736e-191f-4ee8-8820-0ca0efa51ec1 req-b2128c5a-0354-437d-bd7d-aa26f3ecfe61 service nova] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Updated VIF entry in instance network info cache for port 5736bca4-8f2f-485a-bc60-41a19416448c. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 978.007520] env[61868]: DEBUG nova.network.neutron [req-8ae1736e-191f-4ee8-8820-0ca0efa51ec1 req-b2128c5a-0354-437d-bd7d-aa26f3ecfe61 service nova] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Updating instance_info_cache with network_info: [{"id": "5736bca4-8f2f-485a-bc60-41a19416448c", "address": "fa:16:3e:f2:7e:d2", "network": {"id": "f2e3f911-d826-43ad-911d-9f6985db7ee5", "bridge": "br-int", "label": "tempest-ImagesTestJSON-485234144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5658848e1b0e42929c04a4a8de40a291", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5736bca4-8f", "ovs_interfaceid": "5736bca4-8f2f-485a-bc60-41a19416448c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.132923] env[61868]: DEBUG oslo_vmware.api [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e816c2-18b1-39f7-2dac-c8444e32b947, 'name': SearchDatastore_Task, 'duration_secs': 0.053433} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.133297] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.133533] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 978.133775] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.133924] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.134133] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 978.134400] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-796f10a7-7249-44c2-bb54-bde6700363af {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.142914] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 978.143105] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 978.143836] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a85885b1-9949-4f41-8eca-735931081034 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.151285] env[61868]: DEBUG oslo_vmware.api [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 978.151285] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52879019-4acc-0c01-f400-8dab9ef157c5" [ 978.151285] env[61868]: _type = "Task" [ 978.151285] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.158991] env[61868]: DEBUG oslo_vmware.api [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52879019-4acc-0c01-f400-8dab9ef157c5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.229665] env[61868]: DEBUG nova.scheduler.client.report [None req-801169df-e652-4a2f-9396-52c991a74fc1 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 978.233298] env[61868]: DEBUG nova.compute.manager [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 978.245091] env[61868]: DEBUG oslo_vmware.api [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315819, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066199} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.245392] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 978.246240] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a52a9bd-bb83-4ba7-802d-da5f8360c7ff {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.269407] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] a96cab5a-c053-4e45-96f6-2aba0a819110/a96cab5a-c053-4e45-96f6-2aba0a819110.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 978.270268] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8351b170-6057-4e30-b3a1-f99f0579c1d9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.290117] env[61868]: DEBUG oslo_vmware.api [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for the task: (returnval){ [ 978.290117] env[61868]: value = "task-1315820" [ 978.290117] env[61868]: _type = "Task" [ 978.290117] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.298158] env[61868]: DEBUG oslo_vmware.api [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315820, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.388776] env[61868]: DEBUG oslo_vmware.api [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315818, 'name': PowerOnVM_Task, 'duration_secs': 0.863491} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.389169] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 978.389240] env[61868]: INFO nova.compute.manager [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Took 7.93 seconds to spawn the instance on the hypervisor. [ 978.389422] env[61868]: DEBUG nova.compute.manager [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 978.390165] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeced190-56c7-4a41-9fdf-9c74f172cc9e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.511034] env[61868]: DEBUG oslo_concurrency.lockutils [req-8ae1736e-191f-4ee8-8820-0ca0efa51ec1 req-b2128c5a-0354-437d-bd7d-aa26f3ecfe61 service nova] Releasing lock "refresh_cache-cbeeb7b1-166b-49f7-88c5-6ed05c9dc759" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.662007] env[61868]: DEBUG oslo_vmware.api [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52879019-4acc-0c01-f400-8dab9ef157c5, 'name': SearchDatastore_Task, 'duration_secs': 0.01326} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.662884] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa3f085b-6dd7-4256-a2f9-f1acf95ab845 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.670951] env[61868]: DEBUG oslo_vmware.api [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 978.670951] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5246d599-fda9-b19a-4c0f-1d795b51a569" [ 978.670951] env[61868]: _type = "Task" [ 978.670951] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.680060] env[61868]: DEBUG oslo_vmware.api [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5246d599-fda9-b19a-4c0f-1d795b51a569, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.737880] env[61868]: DEBUG oslo_concurrency.lockutils [None req-801169df-e652-4a2f-9396-52c991a74fc1 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.744s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.740651] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.134s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.742521] env[61868]: INFO nova.compute.claims [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 978.761146] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.766881] env[61868]: INFO nova.scheduler.client.report [None req-801169df-e652-4a2f-9396-52c991a74fc1 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Deleted allocations for instance 4216e9b6-7b86-409e-9543-1c57e0f4b850 [ 978.804362] env[61868]: DEBUG oslo_vmware.api [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315820, 'name': ReconfigVM_Task, 'duration_secs': 0.413502} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.804362] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Reconfigured VM instance instance-0000005a to attach disk [datastore1] a96cab5a-c053-4e45-96f6-2aba0a819110/a96cab5a-c053-4e45-96f6-2aba0a819110.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 978.804362] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4eb5211d-e8ec-497d-a6d3-06fcaf59330b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.812091] env[61868]: DEBUG oslo_vmware.api [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for the task: (returnval){ [ 978.812091] env[61868]: value = "task-1315821" [ 978.812091] env[61868]: _type = "Task" [ 978.812091] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.821267] env[61868]: DEBUG oslo_vmware.api [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315821, 'name': Rename_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.911182] env[61868]: INFO nova.compute.manager [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Took 14.91 seconds to build instance. [ 979.184468] env[61868]: DEBUG oslo_vmware.api [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5246d599-fda9-b19a-4c0f-1d795b51a569, 'name': SearchDatastore_Task, 'duration_secs': 0.016794} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.184787] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.185064] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] cbeeb7b1-166b-49f7-88c5-6ed05c9dc759/cbeeb7b1-166b-49f7-88c5-6ed05c9dc759.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 979.185334] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fdd748fb-c2c1-4cb1-8eef-6b776099b977 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.193531] env[61868]: DEBUG oslo_vmware.api [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 979.193531] env[61868]: value = "task-1315822" [ 979.193531] env[61868]: _type = "Task" [ 979.193531] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.203136] env[61868]: DEBUG oslo_vmware.api [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315822, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.274499] env[61868]: DEBUG oslo_concurrency.lockutils [None req-801169df-e652-4a2f-9396-52c991a74fc1 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "4216e9b6-7b86-409e-9543-1c57e0f4b850" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.149s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.321497] env[61868]: DEBUG oslo_vmware.api [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315821, 'name': Rename_Task, 'duration_secs': 0.439492} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.321770] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 979.322013] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bfe1a311-d268-41a0-9db3-e94239457166 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.329913] env[61868]: DEBUG oslo_vmware.api [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for the task: (returnval){ [ 979.329913] env[61868]: value = "task-1315823" [ 979.329913] env[61868]: _type = "Task" [ 979.329913] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.339353] env[61868]: DEBUG oslo_vmware.api [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315823, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.413799] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6e0fc723-38ab-47a7-9047-3867ca6d43c3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "5d2b56d5-a4e0-4dba-9158-71036d60a40b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.427s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.703827] env[61868]: DEBUG oslo_vmware.api [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315822, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.842987] env[61868]: DEBUG oslo_vmware.api [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315823, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.891795] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee435cdd-2925-4512-80cc-10775b08d6f6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.900304] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a024cdd6-dd50-4feb-b8e5-57ec3b741e8c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.932769] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b562a7d8-e051-42dc-903f-2683eb57544b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.946457] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-227969ab-60a7-48dd-93fe-ff27984d02c5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.964020] env[61868]: DEBUG nova.compute.provider_tree [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 980.203451] env[61868]: DEBUG oslo_vmware.api [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315822, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.900769} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.203741] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] cbeeb7b1-166b-49f7-88c5-6ed05c9dc759/cbeeb7b1-166b-49f7-88c5-6ed05c9dc759.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 980.203964] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 980.204237] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0ef00f7f-d1a2-4767-a40a-64d915996e21 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.211323] env[61868]: DEBUG oslo_vmware.api [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 980.211323] env[61868]: value = "task-1315824" [ 980.211323] env[61868]: _type = "Task" [ 980.211323] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.219767] env[61868]: DEBUG oslo_vmware.api [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315824, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.340922] env[61868]: DEBUG oslo_vmware.api [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315823, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.446477] env[61868]: DEBUG nova.compute.manager [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Stashing vm_state: active {{(pid=61868) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 980.469884] env[61868]: DEBUG nova.scheduler.client.report [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 980.566645] env[61868]: DEBUG oslo_concurrency.lockutils [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "54dacd73-c760-49bd-9baf-ad2012a53264" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.566888] env[61868]: DEBUG oslo_concurrency.lockutils [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "54dacd73-c760-49bd-9baf-ad2012a53264" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.721690] env[61868]: DEBUG oslo_vmware.api [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315824, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070273} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.722038] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 980.722957] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f3fa90a-706c-44f0-b5f6-9fa4123b288e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.746547] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] cbeeb7b1-166b-49f7-88c5-6ed05c9dc759/cbeeb7b1-166b-49f7-88c5-6ed05c9dc759.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 980.746870] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7a6f01e7-d75e-4c9d-9298-0134a6d207db {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.766943] env[61868]: DEBUG oslo_vmware.api [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 980.766943] env[61868]: value = "task-1315825" [ 980.766943] env[61868]: _type = "Task" [ 980.766943] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.775472] env[61868]: DEBUG oslo_vmware.api [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315825, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.842740] env[61868]: DEBUG oslo_vmware.api [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315823, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.968141] env[61868]: DEBUG oslo_concurrency.lockutils [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.975129] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.234s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.975662] env[61868]: DEBUG nova.compute.manager [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 980.978862] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.213s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.979158] env[61868]: DEBUG nova.objects.instance [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lazy-loading 'pci_requests' on Instance uuid 75fdbc11-1ef3-442d-bcf4-55069ba9ab62 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 981.069465] env[61868]: DEBUG nova.compute.manager [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 981.277093] env[61868]: DEBUG oslo_vmware.api [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315825, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.341406] env[61868]: DEBUG oslo_vmware.api [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315823, 'name': PowerOnVM_Task, 'duration_secs': 1.593864} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.341718] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 981.341952] env[61868]: INFO nova.compute.manager [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Took 8.56 seconds to spawn the instance on the hypervisor. [ 981.342160] env[61868]: DEBUG nova.compute.manager [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 981.343016] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2720b7dd-79ca-464d-b9b4-b65e48058404 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.481970] env[61868]: DEBUG nova.compute.utils [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 981.484554] env[61868]: DEBUG nova.objects.instance [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lazy-loading 'numa_topology' on Instance uuid 75fdbc11-1ef3-442d-bcf4-55069ba9ab62 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 981.485594] env[61868]: DEBUG nova.compute.manager [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 981.485764] env[61868]: DEBUG nova.network.neutron [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 981.534253] env[61868]: DEBUG nova.policy [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0fa71d6f8941441d89709e7e3af1ed9f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c8440ce785c344d793dd513b5f008d65', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 981.586948] env[61868]: DEBUG oslo_concurrency.lockutils [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.777527] env[61868]: DEBUG oslo_vmware.api [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315825, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.792772] env[61868]: DEBUG nova.network.neutron [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Successfully created port: 8f5f51e1-8b43-4ef6-b9b8-19c6b4e89e6e {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 981.861439] env[61868]: INFO nova.compute.manager [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Took 16.89 seconds to build instance. [ 981.986491] env[61868]: DEBUG nova.compute.manager [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 981.990030] env[61868]: INFO nova.compute.claims [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 982.278026] env[61868]: DEBUG oslo_vmware.api [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315825, 'name': ReconfigVM_Task, 'duration_secs': 1.154849} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.278377] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Reconfigured VM instance instance-0000005b to attach disk [datastore1] cbeeb7b1-166b-49f7-88c5-6ed05c9dc759/cbeeb7b1-166b-49f7-88c5-6ed05c9dc759.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 982.279044] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3d8c56f5-4c8f-4d70-a1f7-3e80845d4ef1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.288199] env[61868]: DEBUG oslo_vmware.api [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 982.288199] env[61868]: value = "task-1315826" [ 982.288199] env[61868]: _type = "Task" [ 982.288199] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.300628] env[61868]: DEBUG oslo_vmware.api [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315826, 'name': Rename_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.362433] env[61868]: DEBUG oslo_concurrency.lockutils [None req-46998913-4bfb-4942-a407-3d628cea6acc tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "a96cab5a-c053-4e45-96f6-2aba0a819110" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.401s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.799273] env[61868]: DEBUG oslo_vmware.api [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315826, 'name': Rename_Task, 'duration_secs': 0.151807} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.799642] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 982.800296] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-53e2b95a-8527-412c-a800-c315e5f44f27 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.810020] env[61868]: DEBUG oslo_vmware.api [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 982.810020] env[61868]: value = "task-1315827" [ 982.810020] env[61868]: _type = "Task" [ 982.810020] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.814106] env[61868]: DEBUG oslo_vmware.api [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315827, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.886054] env[61868]: DEBUG nova.compute.manager [req-8b701802-6a14-4422-bfc3-7e03274341af req-05e02381-2bc4-4acd-b361-636396356f79 service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Received event network-changed-cd9efdc6-e6f5-4449-b95d-d1fbec53a295 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 982.886275] env[61868]: DEBUG nova.compute.manager [req-8b701802-6a14-4422-bfc3-7e03274341af req-05e02381-2bc4-4acd-b361-636396356f79 service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Refreshing instance network info cache due to event network-changed-cd9efdc6-e6f5-4449-b95d-d1fbec53a295. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 982.886643] env[61868]: DEBUG oslo_concurrency.lockutils [req-8b701802-6a14-4422-bfc3-7e03274341af req-05e02381-2bc4-4acd-b361-636396356f79 service nova] Acquiring lock "refresh_cache-a96cab5a-c053-4e45-96f6-2aba0a819110" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.886797] env[61868]: DEBUG oslo_concurrency.lockutils [req-8b701802-6a14-4422-bfc3-7e03274341af req-05e02381-2bc4-4acd-b361-636396356f79 service nova] Acquired lock "refresh_cache-a96cab5a-c053-4e45-96f6-2aba0a819110" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.887009] env[61868]: DEBUG nova.network.neutron [req-8b701802-6a14-4422-bfc3-7e03274341af req-05e02381-2bc4-4acd-b361-636396356f79 service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Refreshing network info cache for port cd9efdc6-e6f5-4449-b95d-d1fbec53a295 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 983.002160] env[61868]: DEBUG nova.compute.manager [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 983.031450] env[61868]: DEBUG nova.virt.hardware [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 983.032081] env[61868]: DEBUG nova.virt.hardware [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 983.032408] env[61868]: DEBUG nova.virt.hardware [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 983.032869] env[61868]: DEBUG nova.virt.hardware [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 983.033157] env[61868]: DEBUG nova.virt.hardware [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 983.033812] env[61868]: DEBUG nova.virt.hardware [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 983.034175] env[61868]: DEBUG nova.virt.hardware [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 983.034825] env[61868]: DEBUG nova.virt.hardware [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 983.038016] env[61868]: DEBUG nova.virt.hardware [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 983.038016] env[61868]: DEBUG nova.virt.hardware [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 983.038016] env[61868]: DEBUG nova.virt.hardware [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 983.038016] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1b798ba-ea91-493b-aadc-a41412335ec0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.052416] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ff6ee54-f45a-424e-962d-1a50c477ff2e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.158722] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b78b7961-bad7-4ade-98ab-37bb3f10a1df {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.166744] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7f33974-1c0b-4238-8cf6-e23d8ba4e3b8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.197929] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f355f956-043e-4885-a36d-30af27ad3e42 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.205545] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7588555f-336f-4abf-ac96-c4fb24b29af8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.224739] env[61868]: DEBUG nova.compute.provider_tree [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 983.270501] env[61868]: DEBUG nova.compute.manager [req-781b5b4e-c315-4b6c-a040-331a8b2d78ee req-869b3d8f-6cca-4ee6-a38b-61dec14b8d13 service nova] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Received event network-vif-plugged-8f5f51e1-8b43-4ef6-b9b8-19c6b4e89e6e {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 983.270732] env[61868]: DEBUG oslo_concurrency.lockutils [req-781b5b4e-c315-4b6c-a040-331a8b2d78ee req-869b3d8f-6cca-4ee6-a38b-61dec14b8d13 service nova] Acquiring lock "eeeec752-7e99-471d-9425-c4ccf4f6d0a5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.271009] env[61868]: DEBUG oslo_concurrency.lockutils [req-781b5b4e-c315-4b6c-a040-331a8b2d78ee req-869b3d8f-6cca-4ee6-a38b-61dec14b8d13 service nova] Lock "eeeec752-7e99-471d-9425-c4ccf4f6d0a5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.271205] env[61868]: DEBUG oslo_concurrency.lockutils [req-781b5b4e-c315-4b6c-a040-331a8b2d78ee req-869b3d8f-6cca-4ee6-a38b-61dec14b8d13 service nova] Lock "eeeec752-7e99-471d-9425-c4ccf4f6d0a5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.271381] env[61868]: DEBUG nova.compute.manager [req-781b5b4e-c315-4b6c-a040-331a8b2d78ee req-869b3d8f-6cca-4ee6-a38b-61dec14b8d13 service nova] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] No waiting events found dispatching network-vif-plugged-8f5f51e1-8b43-4ef6-b9b8-19c6b4e89e6e {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 983.271548] env[61868]: WARNING nova.compute.manager [req-781b5b4e-c315-4b6c-a040-331a8b2d78ee req-869b3d8f-6cca-4ee6-a38b-61dec14b8d13 service nova] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Received unexpected event network-vif-plugged-8f5f51e1-8b43-4ef6-b9b8-19c6b4e89e6e for instance with vm_state building and task_state spawning. [ 983.333843] env[61868]: DEBUG oslo_vmware.api [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315827, 'name': PowerOnVM_Task, 'duration_secs': 0.482921} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.333843] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 983.333843] env[61868]: INFO nova.compute.manager [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Took 8.30 seconds to spawn the instance on the hypervisor. [ 983.333843] env[61868]: DEBUG nova.compute.manager [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 983.333843] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73dc8d96-45a6-431e-adb6-81b001b27533 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.433138] env[61868]: DEBUG nova.network.neutron [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Successfully updated port: 8f5f51e1-8b43-4ef6-b9b8-19c6b4e89e6e {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 983.695499] env[61868]: DEBUG nova.network.neutron [req-8b701802-6a14-4422-bfc3-7e03274341af req-05e02381-2bc4-4acd-b361-636396356f79 service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Updated VIF entry in instance network info cache for port cd9efdc6-e6f5-4449-b95d-d1fbec53a295. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 983.695898] env[61868]: DEBUG nova.network.neutron [req-8b701802-6a14-4422-bfc3-7e03274341af req-05e02381-2bc4-4acd-b361-636396356f79 service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Updating instance_info_cache with network_info: [{"id": "cd9efdc6-e6f5-4449-b95d-d1fbec53a295", "address": "fa:16:3e:bf:13:19", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd9efdc6-e6", "ovs_interfaceid": "cd9efdc6-e6f5-4449-b95d-d1fbec53a295", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.724082] env[61868]: DEBUG nova.scheduler.client.report [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 983.851949] env[61868]: INFO nova.compute.manager [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Took 17.90 seconds to build instance. [ 983.933912] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "refresh_cache-eeeec752-7e99-471d-9425-c4ccf4f6d0a5" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.934080] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquired lock "refresh_cache-eeeec752-7e99-471d-9425-c4ccf4f6d0a5" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.934221] env[61868]: DEBUG nova.network.neutron [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 984.022277] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "69a0b94b-2bf4-49ad-8fb0-ea945176e5eb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.022537] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "69a0b94b-2bf4-49ad-8fb0-ea945176e5eb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.198496] env[61868]: DEBUG oslo_concurrency.lockutils [req-8b701802-6a14-4422-bfc3-7e03274341af req-05e02381-2bc4-4acd-b361-636396356f79 service nova] Releasing lock "refresh_cache-a96cab5a-c053-4e45-96f6-2aba0a819110" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.229802] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.251s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.232401] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.471s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.234012] env[61868]: INFO nova.compute.claims [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 984.279376] env[61868]: INFO nova.network.neutron [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Updating port 21d6cab8-f691-425c-bfa8-9d2f1c294b64 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 984.354688] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a38193bc-9e68-400b-92df-3e5ebc7f36f8 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "cbeeb7b1-166b-49f7-88c5-6ed05c9dc759" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.408s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.468642] env[61868]: DEBUG nova.network.neutron [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 984.524719] env[61868]: DEBUG nova.compute.manager [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 984.635492] env[61868]: DEBUG nova.network.neutron [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Updating instance_info_cache with network_info: [{"id": "8f5f51e1-8b43-4ef6-b9b8-19c6b4e89e6e", "address": "fa:16:3e:88:53:14", "network": {"id": "3535e048-fe57-4cbc-9489-0b77484069bd", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-748518849-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c8440ce785c344d793dd513b5f008d65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f5f51e1-8b", "ovs_interfaceid": "8f5f51e1-8b43-4ef6-b9b8-19c6b4e89e6e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.047307] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.137348] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Releasing lock "refresh_cache-eeeec752-7e99-471d-9425-c4ccf4f6d0a5" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.137740] env[61868]: DEBUG nova.compute.manager [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Instance network_info: |[{"id": "8f5f51e1-8b43-4ef6-b9b8-19c6b4e89e6e", "address": "fa:16:3e:88:53:14", "network": {"id": "3535e048-fe57-4cbc-9489-0b77484069bd", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-748518849-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c8440ce785c344d793dd513b5f008d65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f5f51e1-8b", "ovs_interfaceid": "8f5f51e1-8b43-4ef6-b9b8-19c6b4e89e6e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 985.138227] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:88:53:14', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'da0e5087-d65b-416f-90fe-beaa9c534ad3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8f5f51e1-8b43-4ef6-b9b8-19c6b4e89e6e', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 985.147762] env[61868]: DEBUG oslo.service.loopingcall [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 985.148455] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 985.148808] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9947fd99-f756-40ed-ba0d-acfa5bd9790a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.176745] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 985.176745] env[61868]: value = "task-1315828" [ 985.176745] env[61868]: _type = "Task" [ 985.176745] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.186912] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315828, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.299697] env[61868]: DEBUG nova.compute.manager [req-d58d5464-f2a1-474a-894e-1f7e3cbb4768 req-557d6c4a-3f1b-45a5-b875-b50ae5584c6a service nova] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Received event network-changed-8f5f51e1-8b43-4ef6-b9b8-19c6b4e89e6e {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 985.299939] env[61868]: DEBUG nova.compute.manager [req-d58d5464-f2a1-474a-894e-1f7e3cbb4768 req-557d6c4a-3f1b-45a5-b875-b50ae5584c6a service nova] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Refreshing instance network info cache due to event network-changed-8f5f51e1-8b43-4ef6-b9b8-19c6b4e89e6e. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 985.300159] env[61868]: DEBUG oslo_concurrency.lockutils [req-d58d5464-f2a1-474a-894e-1f7e3cbb4768 req-557d6c4a-3f1b-45a5-b875-b50ae5584c6a service nova] Acquiring lock "refresh_cache-eeeec752-7e99-471d-9425-c4ccf4f6d0a5" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.300292] env[61868]: DEBUG oslo_concurrency.lockutils [req-d58d5464-f2a1-474a-894e-1f7e3cbb4768 req-557d6c4a-3f1b-45a5-b875-b50ae5584c6a service nova] Acquired lock "refresh_cache-eeeec752-7e99-471d-9425-c4ccf4f6d0a5" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.300460] env[61868]: DEBUG nova.network.neutron [req-d58d5464-f2a1-474a-894e-1f7e3cbb4768 req-557d6c4a-3f1b-45a5-b875-b50ae5584c6a service nova] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Refreshing network info cache for port 8f5f51e1-8b43-4ef6-b9b8-19c6b4e89e6e {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 985.382456] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ede6436-b57e-45e8-bf05-bbc6ca2bac72 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.389977] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38f4f4e1-fc88-40bd-bb44-0259dd0b71cd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.419052] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39e2088e-b984-4b93-bf0f-1f8d9e1c6d84 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.426521] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d2ec5e6-0f8b-4bcb-bd79-aefad073f49c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.441853] env[61868]: DEBUG nova.compute.provider_tree [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 985.590916] env[61868]: DEBUG nova.compute.manager [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 985.591473] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7375f37d-b023-4a9c-b1eb-bef1e1cbbf94 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.687018] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315828, 'name': CreateVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.773798] env[61868]: DEBUG nova.compute.manager [req-3c4645c5-0847-435c-a292-1f8a74417484 req-83eb4e3d-1469-46aa-9e95-d577413bd121 service nova] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Received event network-vif-plugged-21d6cab8-f691-425c-bfa8-9d2f1c294b64 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 985.773973] env[61868]: DEBUG oslo_concurrency.lockutils [req-3c4645c5-0847-435c-a292-1f8a74417484 req-83eb4e3d-1469-46aa-9e95-d577413bd121 service nova] Acquiring lock "75fdbc11-1ef3-442d-bcf4-55069ba9ab62-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.774212] env[61868]: DEBUG oslo_concurrency.lockutils [req-3c4645c5-0847-435c-a292-1f8a74417484 req-83eb4e3d-1469-46aa-9e95-d577413bd121 service nova] Lock "75fdbc11-1ef3-442d-bcf4-55069ba9ab62-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.774387] env[61868]: DEBUG oslo_concurrency.lockutils [req-3c4645c5-0847-435c-a292-1f8a74417484 req-83eb4e3d-1469-46aa-9e95-d577413bd121 service nova] Lock "75fdbc11-1ef3-442d-bcf4-55069ba9ab62-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.774564] env[61868]: DEBUG nova.compute.manager [req-3c4645c5-0847-435c-a292-1f8a74417484 req-83eb4e3d-1469-46aa-9e95-d577413bd121 service nova] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] No waiting events found dispatching network-vif-plugged-21d6cab8-f691-425c-bfa8-9d2f1c294b64 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 985.774731] env[61868]: WARNING nova.compute.manager [req-3c4645c5-0847-435c-a292-1f8a74417484 req-83eb4e3d-1469-46aa-9e95-d577413bd121 service nova] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Received unexpected event network-vif-plugged-21d6cab8-f691-425c-bfa8-9d2f1c294b64 for instance with vm_state shelved_offloaded and task_state spawning. [ 985.945188] env[61868]: DEBUG nova.scheduler.client.report [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 985.949211] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquiring lock "refresh_cache-75fdbc11-1ef3-442d-bcf4-55069ba9ab62" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.949360] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquired lock "refresh_cache-75fdbc11-1ef3-442d-bcf4-55069ba9ab62" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.949541] env[61868]: DEBUG nova.network.neutron [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 986.018968] env[61868]: DEBUG nova.network.neutron [req-d58d5464-f2a1-474a-894e-1f7e3cbb4768 req-557d6c4a-3f1b-45a5-b875-b50ae5584c6a service nova] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Updated VIF entry in instance network info cache for port 8f5f51e1-8b43-4ef6-b9b8-19c6b4e89e6e. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 986.019317] env[61868]: DEBUG nova.network.neutron [req-d58d5464-f2a1-474a-894e-1f7e3cbb4768 req-557d6c4a-3f1b-45a5-b875-b50ae5584c6a service nova] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Updating instance_info_cache with network_info: [{"id": "8f5f51e1-8b43-4ef6-b9b8-19c6b4e89e6e", "address": "fa:16:3e:88:53:14", "network": {"id": "3535e048-fe57-4cbc-9489-0b77484069bd", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-748518849-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c8440ce785c344d793dd513b5f008d65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f5f51e1-8b", "ovs_interfaceid": "8f5f51e1-8b43-4ef6-b9b8-19c6b4e89e6e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.103531] env[61868]: INFO nova.compute.manager [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] instance snapshotting [ 986.107396] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ba4ad0f-324b-4829-9e2d-2642521b3220 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.127231] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4394f422-de8c-4777-b5d4-06e2dd38fc76 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.187977] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315828, 'name': CreateVM_Task, 'duration_secs': 0.56403} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.188160] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 986.188772] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.188941] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.189277] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 986.189519] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e796213e-57e5-4ff7-b1f1-d3de12bc245d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.194024] env[61868]: DEBUG oslo_vmware.api [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 986.194024] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52f7ffea-0a8c-6782-221c-52e9ae563147" [ 986.194024] env[61868]: _type = "Task" [ 986.194024] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.201657] env[61868]: DEBUG oslo_vmware.api [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52f7ffea-0a8c-6782-221c-52e9ae563147, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.457564] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.223s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.457564] env[61868]: DEBUG nova.compute.manager [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 986.460921] env[61868]: DEBUG oslo_concurrency.lockutils [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 5.492s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.522314] env[61868]: DEBUG oslo_concurrency.lockutils [req-d58d5464-f2a1-474a-894e-1f7e3cbb4768 req-557d6c4a-3f1b-45a5-b875-b50ae5584c6a service nova] Releasing lock "refresh_cache-eeeec752-7e99-471d-9425-c4ccf4f6d0a5" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.637734] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Creating Snapshot of the VM instance {{(pid=61868) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 986.638063] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-bbed7f5a-4193-4dec-94bd-0105fcbeecff {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.645157] env[61868]: DEBUG nova.network.neutron [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Updating instance_info_cache with network_info: [{"id": "21d6cab8-f691-425c-bfa8-9d2f1c294b64", "address": "fa:16:3e:76:a4:70", "network": {"id": "e7457f1a-171a-4c82-a260-1d1a7a95861e", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1032778554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b9b800086764b41a8817d1d8e463a2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21d6cab8-f6", "ovs_interfaceid": "21d6cab8-f691-425c-bfa8-9d2f1c294b64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.647433] env[61868]: DEBUG oslo_vmware.api [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 986.647433] env[61868]: value = "task-1315829" [ 986.647433] env[61868]: _type = "Task" [ 986.647433] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.655555] env[61868]: DEBUG oslo_vmware.api [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315829, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.704128] env[61868]: DEBUG oslo_vmware.api [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52f7ffea-0a8c-6782-221c-52e9ae563147, 'name': SearchDatastore_Task, 'duration_secs': 0.010336} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.704448] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.704701] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 986.704937] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.705100] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.705284] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 986.705580] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e52a627f-b9af-430c-9cde-dec8052c94b7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.713536] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 986.713717] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 986.714507] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe8e690b-eeba-42d5-926d-c3a75f13a1f6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.719551] env[61868]: DEBUG oslo_vmware.api [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 986.719551] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52056ebd-551f-9630-16d1-3c9cc4f622ab" [ 986.719551] env[61868]: _type = "Task" [ 986.719551] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.727747] env[61868]: DEBUG oslo_vmware.api [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52056ebd-551f-9630-16d1-3c9cc4f622ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.962330] env[61868]: DEBUG nova.compute.utils [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 986.964775] env[61868]: DEBUG nova.compute.manager [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 986.964775] env[61868]: DEBUG nova.network.neutron [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 986.968940] env[61868]: INFO nova.compute.claims [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 987.013518] env[61868]: DEBUG nova.policy [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'daad105e7edf4fb0ae0b2e685bfd92e9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b520c3ac58074e8d9b0bfafb817244a8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 987.148921] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Releasing lock "refresh_cache-75fdbc11-1ef3-442d-bcf4-55069ba9ab62" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.161352] env[61868]: DEBUG oslo_vmware.api [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315829, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.173819] env[61868]: DEBUG nova.virt.hardware [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='2c6a26eb420de28ad4ef204412527bcb',container_format='bare',created_at=2024-10-18T17:25:57Z,direct_url=,disk_format='vmdk',id=76af0b0a-47a5-487a-9b10-d16668b6caae,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-1164793827-shelved',owner='2b9b800086764b41a8817d1d8e463a2f',properties=ImageMetaProps,protected=,size=31665664,status='active',tags=,updated_at=2024-10-18T17:26:10Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 987.174089] env[61868]: DEBUG nova.virt.hardware [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 987.174255] env[61868]: DEBUG nova.virt.hardware [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 987.174451] env[61868]: DEBUG nova.virt.hardware [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 987.174611] env[61868]: DEBUG nova.virt.hardware [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 987.175071] env[61868]: DEBUG nova.virt.hardware [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 987.175071] env[61868]: DEBUG nova.virt.hardware [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 987.175201] env[61868]: DEBUG nova.virt.hardware [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 987.175330] env[61868]: DEBUG nova.virt.hardware [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 987.175501] env[61868]: DEBUG nova.virt.hardware [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 987.175677] env[61868]: DEBUG nova.virt.hardware [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 987.176599] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f962fc2-76ea-4e5b-b259-015b463f5113 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.185128] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e429d5d-9a0f-40f6-b236-8356efdc605f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.204218] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:76:a4:70', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd19577c9-1b2e-490b-8031-2f278dd3f570', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '21d6cab8-f691-425c-bfa8-9d2f1c294b64', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 987.211878] env[61868]: DEBUG oslo.service.loopingcall [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 987.212160] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 987.212403] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7f436823-e998-4fa5-8cd0-4eee67fd4a0b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.237404] env[61868]: DEBUG oslo_vmware.api [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52056ebd-551f-9630-16d1-3c9cc4f622ab, 'name': SearchDatastore_Task, 'duration_secs': 0.009897} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.239445] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 987.239445] env[61868]: value = "task-1315830" [ 987.239445] env[61868]: _type = "Task" [ 987.239445] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.239689] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e22a020-e8e8-4085-9572-027dccf446c5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.251069] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315830, 'name': CreateVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.252381] env[61868]: DEBUG oslo_vmware.api [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 987.252381] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52a6526e-6d3c-25b8-93ef-bbb109205c6e" [ 987.252381] env[61868]: _type = "Task" [ 987.252381] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.262485] env[61868]: DEBUG oslo_vmware.api [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52a6526e-6d3c-25b8-93ef-bbb109205c6e, 'name': SearchDatastore_Task, 'duration_secs': 0.01013} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.265690] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.265690] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] eeeec752-7e99-471d-9425-c4ccf4f6d0a5/eeeec752-7e99-471d-9425-c4ccf4f6d0a5.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 987.265690] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3139ae91-85a8-4f83-8ac9-1df4fdc571fb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.279672] env[61868]: DEBUG oslo_vmware.api [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 987.279672] env[61868]: value = "task-1315831" [ 987.279672] env[61868]: _type = "Task" [ 987.279672] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.290108] env[61868]: DEBUG oslo_vmware.api [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315831, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.303634] env[61868]: DEBUG nova.network.neutron [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Successfully created port: 80d0494a-2f37-48c5-b7c6-9491f5da2bc4 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 987.474016] env[61868]: DEBUG nova.compute.manager [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 987.479813] env[61868]: INFO nova.compute.resource_tracker [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Updating resource usage from migration 9f4d2d6e-1376-4d88-8eb6-b575d737488f [ 987.616649] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 987.616989] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 987.664430] env[61868]: DEBUG oslo_vmware.api [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315829, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.683501] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5451b06-cb7c-4a96-b9e5-27c142001b11 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.694068] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3cab235-10df-4eea-bde9-e44ec6b841f9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.731084] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3846bed9-10c8-45fd-9f60-c5d9ba485963 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.739703] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80f27e66-b0df-43d1-a13f-3c7e20d4502a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.757798] env[61868]: DEBUG nova.compute.provider_tree [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 987.762966] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315830, 'name': CreateVM_Task, 'duration_secs': 0.392616} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.764326] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 987.764326] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/76af0b0a-47a5-487a-9b10-d16668b6caae" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.764326] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquired lock "[datastore1] devstack-image-cache_base/76af0b0a-47a5-487a-9b10-d16668b6caae" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.764665] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/76af0b0a-47a5-487a-9b10-d16668b6caae" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 987.765015] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d52a1b1-d181-43d6-9a77-43d9ee234854 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.773342] env[61868]: DEBUG oslo_vmware.api [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the task: (returnval){ [ 987.773342] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]528e912a-9b4a-7a7a-be00-aa9af57e2e2a" [ 987.773342] env[61868]: _type = "Task" [ 987.773342] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.787817] env[61868]: DEBUG oslo_vmware.api [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]528e912a-9b4a-7a7a-be00-aa9af57e2e2a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.793135] env[61868]: DEBUG oslo_vmware.api [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315831, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.811551] env[61868]: DEBUG nova.compute.manager [req-034e9b6c-114c-4f3b-9ecb-31202cf96183 req-b77c402d-ce54-48ce-a545-f466efd813a0 service nova] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Received event network-changed-21d6cab8-f691-425c-bfa8-9d2f1c294b64 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 987.811802] env[61868]: DEBUG nova.compute.manager [req-034e9b6c-114c-4f3b-9ecb-31202cf96183 req-b77c402d-ce54-48ce-a545-f466efd813a0 service nova] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Refreshing instance network info cache due to event network-changed-21d6cab8-f691-425c-bfa8-9d2f1c294b64. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 987.812148] env[61868]: DEBUG oslo_concurrency.lockutils [req-034e9b6c-114c-4f3b-9ecb-31202cf96183 req-b77c402d-ce54-48ce-a545-f466efd813a0 service nova] Acquiring lock "refresh_cache-75fdbc11-1ef3-442d-bcf4-55069ba9ab62" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.812401] env[61868]: DEBUG oslo_concurrency.lockutils [req-034e9b6c-114c-4f3b-9ecb-31202cf96183 req-b77c402d-ce54-48ce-a545-f466efd813a0 service nova] Acquired lock "refresh_cache-75fdbc11-1ef3-442d-bcf4-55069ba9ab62" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.812605] env[61868]: DEBUG nova.network.neutron [req-034e9b6c-114c-4f3b-9ecb-31202cf96183 req-b77c402d-ce54-48ce-a545-f466efd813a0 service nova] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Refreshing network info cache for port 21d6cab8-f691-425c-bfa8-9d2f1c294b64 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 988.123844] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 988.124228] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Starting heal instance info cache {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 988.161795] env[61868]: DEBUG oslo_vmware.api [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315829, 'name': CreateSnapshot_Task, 'duration_secs': 1.330722} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.162765] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Created Snapshot of the VM instance {{(pid=61868) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 988.163579] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b782f6b8-6a17-4636-8f0d-a888a1a6cb76 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.264865] env[61868]: DEBUG nova.scheduler.client.report [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 988.288359] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Releasing lock "[datastore1] devstack-image-cache_base/76af0b0a-47a5-487a-9b10-d16668b6caae" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.288633] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Processing image 76af0b0a-47a5-487a-9b10-d16668b6caae {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 988.288853] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/76af0b0a-47a5-487a-9b10-d16668b6caae/76af0b0a-47a5-487a-9b10-d16668b6caae.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.289008] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquired lock "[datastore1] devstack-image-cache_base/76af0b0a-47a5-487a-9b10-d16668b6caae/76af0b0a-47a5-487a-9b10-d16668b6caae.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.289210] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 988.289764] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cb0f843b-bddd-48b9-a245-3371c634ae9c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.297686] env[61868]: DEBUG oslo_vmware.api [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315831, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.527771} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.298757] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] eeeec752-7e99-471d-9425-c4ccf4f6d0a5/eeeec752-7e99-471d-9425-c4ccf4f6d0a5.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 988.298913] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 988.299184] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 988.299349] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 988.300054] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2cba8013-ba0f-4ebe-9b16-a8991c69ad56 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.301781] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6180cad5-9ee7-47c2-8bbf-c1520ed0928c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.307824] env[61868]: DEBUG oslo_vmware.api [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the task: (returnval){ [ 988.307824] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52c4c512-cc43-7aab-409b-02d8c2ad8d0c" [ 988.307824] env[61868]: _type = "Task" [ 988.307824] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.311782] env[61868]: DEBUG oslo_vmware.api [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 988.311782] env[61868]: value = "task-1315832" [ 988.311782] env[61868]: _type = "Task" [ 988.311782] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.319758] env[61868]: DEBUG oslo_vmware.api [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52c4c512-cc43-7aab-409b-02d8c2ad8d0c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.329241] env[61868]: DEBUG oslo_vmware.api [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315832, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.491477] env[61868]: DEBUG nova.compute.manager [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 988.513216] env[61868]: DEBUG nova.virt.hardware [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 988.513477] env[61868]: DEBUG nova.virt.hardware [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 988.513640] env[61868]: DEBUG nova.virt.hardware [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 988.513826] env[61868]: DEBUG nova.virt.hardware [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 988.513974] env[61868]: DEBUG nova.virt.hardware [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 988.514139] env[61868]: DEBUG nova.virt.hardware [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 988.514353] env[61868]: DEBUG nova.virt.hardware [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 988.514514] env[61868]: DEBUG nova.virt.hardware [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 988.514687] env[61868]: DEBUG nova.virt.hardware [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 988.514845] env[61868]: DEBUG nova.virt.hardware [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 988.515036] env[61868]: DEBUG nova.virt.hardware [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 988.515916] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27a41094-f33f-4292-a8a0-e776c5845795 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.526343] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a12f7e90-3206-4131-9803-37105904c477 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.555706] env[61868]: DEBUG nova.network.neutron [req-034e9b6c-114c-4f3b-9ecb-31202cf96183 req-b77c402d-ce54-48ce-a545-f466efd813a0 service nova] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Updated VIF entry in instance network info cache for port 21d6cab8-f691-425c-bfa8-9d2f1c294b64. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 988.556085] env[61868]: DEBUG nova.network.neutron [req-034e9b6c-114c-4f3b-9ecb-31202cf96183 req-b77c402d-ce54-48ce-a545-f466efd813a0 service nova] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Updating instance_info_cache with network_info: [{"id": "21d6cab8-f691-425c-bfa8-9d2f1c294b64", "address": "fa:16:3e:76:a4:70", "network": {"id": "e7457f1a-171a-4c82-a260-1d1a7a95861e", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1032778554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b9b800086764b41a8817d1d8e463a2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21d6cab8-f6", "ovs_interfaceid": "21d6cab8-f691-425c-bfa8-9d2f1c294b64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.682385] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Creating linked-clone VM from snapshot {{(pid=61868) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 988.682740] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-ac4ad9e4-3f14-4878-b288-8eb926661ebc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.690840] env[61868]: DEBUG oslo_vmware.api [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 988.690840] env[61868]: value = "task-1315833" [ 988.690840] env[61868]: _type = "Task" [ 988.690840] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.699155] env[61868]: DEBUG oslo_vmware.api [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315833, 'name': CloneVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.772083] env[61868]: DEBUG oslo_concurrency.lockutils [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.310s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.772083] env[61868]: INFO nova.compute.manager [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Migrating [ 988.784152] env[61868]: DEBUG oslo_concurrency.lockutils [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.197s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.786897] env[61868]: INFO nova.compute.claims [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 988.827050] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Preparing fetch location {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 988.827050] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Fetch image to [datastore1] OSTACK_IMG_db157afc-0490-45e2-b788-b5007ddcbe1b/OSTACK_IMG_db157afc-0490-45e2-b788-b5007ddcbe1b.vmdk {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 988.827515] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Downloading stream optimized image 76af0b0a-47a5-487a-9b10-d16668b6caae to [datastore1] OSTACK_IMG_db157afc-0490-45e2-b788-b5007ddcbe1b/OSTACK_IMG_db157afc-0490-45e2-b788-b5007ddcbe1b.vmdk on the data store datastore1 as vApp {{(pid=61868) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 988.827515] env[61868]: DEBUG nova.virt.vmwareapi.images [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Downloading image file data 76af0b0a-47a5-487a-9b10-d16668b6caae to the ESX as VM named 'OSTACK_IMG_db157afc-0490-45e2-b788-b5007ddcbe1b' {{(pid=61868) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 988.830037] env[61868]: DEBUG oslo_vmware.api [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315832, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.182661} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.831391] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 988.832410] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a53f868-d047-4d85-b7b0-f185e12b88bb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.861936] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] eeeec752-7e99-471d-9425-c4ccf4f6d0a5/eeeec752-7e99-471d-9425-c4ccf4f6d0a5.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 988.867125] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d03f3143-1630-4707-9de8-7f8baf78c4df {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.892431] env[61868]: DEBUG nova.compute.manager [req-151fc6b5-fc65-470f-9a7c-ce01672df59c req-be21231b-686e-4e3a-a521-8bce08354110 service nova] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Received event network-vif-plugged-80d0494a-2f37-48c5-b7c6-9491f5da2bc4 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 988.892850] env[61868]: DEBUG oslo_concurrency.lockutils [req-151fc6b5-fc65-470f-9a7c-ce01672df59c req-be21231b-686e-4e3a-a521-8bce08354110 service nova] Acquiring lock "56703761-3421-44fd-aa63-9db36c86e69c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.893248] env[61868]: DEBUG oslo_concurrency.lockutils [req-151fc6b5-fc65-470f-9a7c-ce01672df59c req-be21231b-686e-4e3a-a521-8bce08354110 service nova] Lock "56703761-3421-44fd-aa63-9db36c86e69c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.893582] env[61868]: DEBUG oslo_concurrency.lockutils [req-151fc6b5-fc65-470f-9a7c-ce01672df59c req-be21231b-686e-4e3a-a521-8bce08354110 service nova] Lock "56703761-3421-44fd-aa63-9db36c86e69c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.893916] env[61868]: DEBUG nova.compute.manager [req-151fc6b5-fc65-470f-9a7c-ce01672df59c req-be21231b-686e-4e3a-a521-8bce08354110 service nova] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] No waiting events found dispatching network-vif-plugged-80d0494a-2f37-48c5-b7c6-9491f5da2bc4 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 988.894277] env[61868]: WARNING nova.compute.manager [req-151fc6b5-fc65-470f-9a7c-ce01672df59c req-be21231b-686e-4e3a-a521-8bce08354110 service nova] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Received unexpected event network-vif-plugged-80d0494a-2f37-48c5-b7c6-9491f5da2bc4 for instance with vm_state building and task_state spawning. [ 988.896848] env[61868]: DEBUG nova.network.neutron [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Successfully updated port: 80d0494a-2f37-48c5-b7c6-9491f5da2bc4 {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 988.921653] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "refresh_cache-56703761-3421-44fd-aa63-9db36c86e69c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.921912] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquired lock "refresh_cache-56703761-3421-44fd-aa63-9db36c86e69c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.922155] env[61868]: DEBUG nova.network.neutron [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 988.926255] env[61868]: DEBUG oslo_vmware.api [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 988.926255] env[61868]: value = "task-1315834" [ 988.926255] env[61868]: _type = "Task" [ 988.926255] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.938824] env[61868]: DEBUG oslo_vmware.api [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315834, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.971042] env[61868]: DEBUG oslo_vmware.rw_handles [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 988.971042] env[61868]: value = "resgroup-9" [ 988.971042] env[61868]: _type = "ResourcePool" [ 988.971042] env[61868]: }. {{(pid=61868) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 988.971520] env[61868]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-7abf968e-6751-43c6-b15b-1751ce33bc6e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.995556] env[61868]: DEBUG oslo_vmware.rw_handles [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lease: (returnval){ [ 988.995556] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52108059-c539-7634-5e9a-c5edc29b59fb" [ 988.995556] env[61868]: _type = "HttpNfcLease" [ 988.995556] env[61868]: } obtained for vApp import into resource pool (val){ [ 988.995556] env[61868]: value = "resgroup-9" [ 988.995556] env[61868]: _type = "ResourcePool" [ 988.995556] env[61868]: }. {{(pid=61868) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 988.995556] env[61868]: DEBUG oslo_vmware.api [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the lease: (returnval){ [ 988.995556] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52108059-c539-7634-5e9a-c5edc29b59fb" [ 988.995556] env[61868]: _type = "HttpNfcLease" [ 988.995556] env[61868]: } to be ready. {{(pid=61868) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 989.001811] env[61868]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 989.001811] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52108059-c539-7634-5e9a-c5edc29b59fb" [ 989.001811] env[61868]: _type = "HttpNfcLease" [ 989.001811] env[61868]: } is initializing. {{(pid=61868) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 989.058856] env[61868]: DEBUG oslo_concurrency.lockutils [req-034e9b6c-114c-4f3b-9ecb-31202cf96183 req-b77c402d-ce54-48ce-a545-f466efd813a0 service nova] Releasing lock "refresh_cache-75fdbc11-1ef3-442d-bcf4-55069ba9ab62" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.201334] env[61868]: DEBUG oslo_vmware.api [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315833, 'name': CloneVM_Task} progress is 94%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.295809] env[61868]: DEBUG oslo_concurrency.lockutils [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "refresh_cache-5d2b56d5-a4e0-4dba-9158-71036d60a40b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.296023] env[61868]: DEBUG oslo_concurrency.lockutils [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquired lock "refresh_cache-5d2b56d5-a4e0-4dba-9158-71036d60a40b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.296164] env[61868]: DEBUG nova.network.neutron [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 989.440613] env[61868]: DEBUG oslo_vmware.api [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315834, 'name': ReconfigVM_Task, 'duration_secs': 0.43391} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.440907] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Reconfigured VM instance instance-0000005c to attach disk [datastore1] eeeec752-7e99-471d-9425-c4ccf4f6d0a5/eeeec752-7e99-471d-9425-c4ccf4f6d0a5.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 989.441569] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-29f59260-d83e-4465-bb34-9f2da77f4ce7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.448061] env[61868]: DEBUG oslo_vmware.api [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 989.448061] env[61868]: value = "task-1315836" [ 989.448061] env[61868]: _type = "Task" [ 989.448061] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.455845] env[61868]: DEBUG oslo_vmware.api [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315836, 'name': Rename_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.465401] env[61868]: DEBUG nova.network.neutron [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 989.503591] env[61868]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 989.503591] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52108059-c539-7634-5e9a-c5edc29b59fb" [ 989.503591] env[61868]: _type = "HttpNfcLease" [ 989.503591] env[61868]: } is initializing. {{(pid=61868) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 989.605491] env[61868]: DEBUG nova.network.neutron [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Updating instance_info_cache with network_info: [{"id": "80d0494a-2f37-48c5-b7c6-9491f5da2bc4", "address": "fa:16:3e:25:77:13", "network": {"id": "77f33b52-e4eb-4b72-85d4-5991a2b5c36d", "bridge": "br-int", "label": "tempest-ServersTestJSON-748635995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b520c3ac58074e8d9b0bfafb817244a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap80d0494a-2f", "ovs_interfaceid": "80d0494a-2f37-48c5-b7c6-9491f5da2bc4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.701421] env[61868]: DEBUG oslo_vmware.api [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315833, 'name': CloneVM_Task} progress is 94%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.944748] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfcb55c4-6fae-4329-b47e-a7932ce6164a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.955177] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aebca7c-0a67-4cc9-ad23-34f2d129cbbd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.961012] env[61868]: DEBUG oslo_vmware.api [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315836, 'name': Rename_Task, 'duration_secs': 0.150278} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.961623] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 989.961881] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-af20d37f-c15b-4e5a-9629-4df452416112 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.993947] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a27b7a8-0038-42bd-b90d-42d70653b1f5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.996490] env[61868]: DEBUG oslo_vmware.api [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 989.996490] env[61868]: value = "task-1315837" [ 989.996490] env[61868]: _type = "Task" [ 989.996490] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.006423] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc655446-cfe0-40bb-93a7-44761815c140 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.014111] env[61868]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 990.014111] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52108059-c539-7634-5e9a-c5edc29b59fb" [ 990.014111] env[61868]: _type = "HttpNfcLease" [ 990.014111] env[61868]: } is ready. {{(pid=61868) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 990.014405] env[61868]: DEBUG oslo_vmware.api [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315837, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.014999] env[61868]: DEBUG oslo_vmware.rw_handles [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 990.014999] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52108059-c539-7634-5e9a-c5edc29b59fb" [ 990.014999] env[61868]: _type = "HttpNfcLease" [ 990.014999] env[61868]: }. {{(pid=61868) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 990.015697] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1131731-7680-4bab-b7ad-fe79f1c09367 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.028185] env[61868]: DEBUG nova.compute.provider_tree [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 990.036234] env[61868]: DEBUG oslo_vmware.rw_handles [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52795126-451e-14d3-fb9f-527bc0f8b200/disk-0.vmdk from lease info. {{(pid=61868) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 990.036234] env[61868]: DEBUG oslo_vmware.rw_handles [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Creating HTTP connection to write to file with size = 31665664 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52795126-451e-14d3-fb9f-527bc0f8b200/disk-0.vmdk. {{(pid=61868) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 990.039065] env[61868]: DEBUG nova.network.neutron [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Updating instance_info_cache with network_info: [{"id": "9023a9b8-ad2a-42b6-b2cd-b2cdb86dd9cb", "address": "fa:16:3e:be:bc:fe", "network": {"id": "c1f5bf03-438f-4acc-93fd-159bed7d4069", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2061742146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a60c3c20950e4156b64c3b4c61b9f0f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24b91e3b-50f1-4a16-b929-942e6b31b2d8", "external-id": "nsx-vlan-transportzone-34", "segmentation_id": 34, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9023a9b8-ad", "ovs_interfaceid": "9023a9b8-ad2a-42b6-b2cd-b2cdb86dd9cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.095604] env[61868]: DEBUG oslo_concurrency.lockutils [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Releasing lock "refresh_cache-5d2b56d5-a4e0-4dba-9158-71036d60a40b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.102679] env[61868]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-df495660-6c5a-4d0e-bf13-34b061860d66 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.109339] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Releasing lock "refresh_cache-56703761-3421-44fd-aa63-9db36c86e69c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.109615] env[61868]: DEBUG nova.compute.manager [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Instance network_info: |[{"id": "80d0494a-2f37-48c5-b7c6-9491f5da2bc4", "address": "fa:16:3e:25:77:13", "network": {"id": "77f33b52-e4eb-4b72-85d4-5991a2b5c36d", "bridge": "br-int", "label": "tempest-ServersTestJSON-748635995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b520c3ac58074e8d9b0bfafb817244a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap80d0494a-2f", "ovs_interfaceid": "80d0494a-2f37-48c5-b7c6-9491f5da2bc4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 990.110138] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:25:77:13', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56136ef6-99d7-4562-9a9f-d66fec951c5c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '80d0494a-2f37-48c5-b7c6-9491f5da2bc4', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 990.119824] env[61868]: DEBUG oslo.service.loopingcall [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 990.121812] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 990.122065] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-87dad13a-19cf-48a2-b539-165cfd2e31ce {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.145085] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 990.145085] env[61868]: value = "task-1315838" [ 990.145085] env[61868]: _type = "Task" [ 990.145085] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.153325] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315838, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.202774] env[61868]: DEBUG oslo_vmware.api [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315833, 'name': CloneVM_Task} progress is 95%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.506904] env[61868]: DEBUG oslo_vmware.api [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315837, 'name': PowerOnVM_Task, 'duration_secs': 0.526333} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.507335] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 990.507627] env[61868]: INFO nova.compute.manager [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Took 7.51 seconds to spawn the instance on the hypervisor. [ 990.507918] env[61868]: DEBUG nova.compute.manager [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 990.508816] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59ccbb48-f149-42d1-89d4-aeead7c78d2d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.531954] env[61868]: DEBUG nova.scheduler.client.report [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 990.657679] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315838, 'name': CreateVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.704584] env[61868]: DEBUG oslo_vmware.api [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315833, 'name': CloneVM_Task, 'duration_secs': 1.847921} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.706116] env[61868]: INFO nova.virt.vmwareapi.vmops [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Created linked-clone VM from snapshot [ 990.706906] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-157cafa0-cc45-4e03-a528-17eca07d4b2e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.716347] env[61868]: DEBUG nova.virt.vmwareapi.images [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Uploading image 6732cc60-80ef-44f3-89ae-468b81ae48f9 {{(pid=61868) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 990.745872] env[61868]: DEBUG oslo_vmware.rw_handles [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 990.745872] env[61868]: value = "vm-281604" [ 990.745872] env[61868]: _type = "VirtualMachine" [ 990.745872] env[61868]: }. {{(pid=61868) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 990.746201] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-c28633f7-1069-4a86-bbc8-b7d650cfa236 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.753338] env[61868]: DEBUG oslo_vmware.rw_handles [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lease: (returnval){ [ 990.753338] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]529e473a-d9a9-23d9-ecb0-cea0236e1cfa" [ 990.753338] env[61868]: _type = "HttpNfcLease" [ 990.753338] env[61868]: } obtained for exporting VM: (result){ [ 990.753338] env[61868]: value = "vm-281604" [ 990.753338] env[61868]: _type = "VirtualMachine" [ 990.753338] env[61868]: }. {{(pid=61868) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 990.753767] env[61868]: DEBUG oslo_vmware.api [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the lease: (returnval){ [ 990.753767] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]529e473a-d9a9-23d9-ecb0-cea0236e1cfa" [ 990.753767] env[61868]: _type = "HttpNfcLease" [ 990.753767] env[61868]: } to be ready. {{(pid=61868) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 990.760967] env[61868]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 990.760967] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]529e473a-d9a9-23d9-ecb0-cea0236e1cfa" [ 990.760967] env[61868]: _type = "HttpNfcLease" [ 990.760967] env[61868]: } is initializing. {{(pid=61868) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 990.869403] env[61868]: DEBUG nova.compute.manager [req-5c16cf77-c7cc-4bf3-9654-01fd884f443e req-b77ce442-c11f-488a-b33e-a801657fed92 service nova] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Received event network-changed-80d0494a-2f37-48c5-b7c6-9491f5da2bc4 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 990.869403] env[61868]: DEBUG nova.compute.manager [req-5c16cf77-c7cc-4bf3-9654-01fd884f443e req-b77ce442-c11f-488a-b33e-a801657fed92 service nova] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Refreshing instance network info cache due to event network-changed-80d0494a-2f37-48c5-b7c6-9491f5da2bc4. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 990.869403] env[61868]: DEBUG oslo_concurrency.lockutils [req-5c16cf77-c7cc-4bf3-9654-01fd884f443e req-b77ce442-c11f-488a-b33e-a801657fed92 service nova] Acquiring lock "refresh_cache-56703761-3421-44fd-aa63-9db36c86e69c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.869403] env[61868]: DEBUG oslo_concurrency.lockutils [req-5c16cf77-c7cc-4bf3-9654-01fd884f443e req-b77ce442-c11f-488a-b33e-a801657fed92 service nova] Acquired lock "refresh_cache-56703761-3421-44fd-aa63-9db36c86e69c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.869403] env[61868]: DEBUG nova.network.neutron [req-5c16cf77-c7cc-4bf3-9654-01fd884f443e req-b77ce442-c11f-488a-b33e-a801657fed92 service nova] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Refreshing network info cache for port 80d0494a-2f37-48c5-b7c6-9491f5da2bc4 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 991.030040] env[61868]: INFO nova.compute.manager [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Took 19.45 seconds to build instance. [ 991.037800] env[61868]: DEBUG oslo_concurrency.lockutils [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.254s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.038358] env[61868]: DEBUG nova.compute.manager [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 991.042405] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.995s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.043961] env[61868]: INFO nova.compute.claims [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 991.158064] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315838, 'name': CreateVM_Task, 'duration_secs': 0.55185} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.158983] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 991.159797] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.160228] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.160639] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 991.162526] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3c5504f8-cd5e-49e3-bad0-6d3b45334868 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.168081] env[61868]: DEBUG oslo_vmware.api [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 991.168081] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52f61d99-f642-7817-199f-718df4f164ae" [ 991.168081] env[61868]: _type = "Task" [ 991.168081] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.178469] env[61868]: DEBUG oslo_vmware.api [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52f61d99-f642-7817-199f-718df4f164ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.263931] env[61868]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 991.263931] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]529e473a-d9a9-23d9-ecb0-cea0236e1cfa" [ 991.263931] env[61868]: _type = "HttpNfcLease" [ 991.263931] env[61868]: } is ready. {{(pid=61868) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 991.264515] env[61868]: DEBUG oslo_vmware.rw_handles [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 991.264515] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]529e473a-d9a9-23d9-ecb0-cea0236e1cfa" [ 991.264515] env[61868]: _type = "HttpNfcLease" [ 991.264515] env[61868]: }. {{(pid=61868) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 991.267929] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82a474b8-e0e0-44bf-91ec-ea16a9e8a35f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.275186] env[61868]: DEBUG oslo_vmware.rw_handles [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52dc8fad-8f05-f77b-148c-f74b46174d8f/disk-0.vmdk from lease info. {{(pid=61868) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 991.275389] env[61868]: DEBUG oslo_vmware.rw_handles [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52dc8fad-8f05-f77b-148c-f74b46174d8f/disk-0.vmdk for reading. {{(pid=61868) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 991.380036] env[61868]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-bc3da1a8-d9f5-4c20-8d5b-2ed8bc5e23bb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.410124] env[61868]: DEBUG oslo_vmware.rw_handles [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Completed reading data from the image iterator. {{(pid=61868) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 991.410274] env[61868]: DEBUG oslo_vmware.rw_handles [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52795126-451e-14d3-fb9f-527bc0f8b200/disk-0.vmdk. {{(pid=61868) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 991.411153] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6774ae08-5d28-4c8e-a583-473a35368726 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.424429] env[61868]: DEBUG oslo_vmware.rw_handles [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52795126-451e-14d3-fb9f-527bc0f8b200/disk-0.vmdk is in state: ready. {{(pid=61868) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 991.424678] env[61868]: DEBUG oslo_vmware.rw_handles [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52795126-451e-14d3-fb9f-527bc0f8b200/disk-0.vmdk. {{(pid=61868) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 991.425064] env[61868]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-e6051585-a691-4789-a55e-e791de44d8ff {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.532328] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b0f6032a-17f7-4fca-a30e-96ac201ced6e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "eeeec752-7e99-471d-9425-c4ccf4f6d0a5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.959s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.549190] env[61868]: DEBUG nova.compute.utils [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 991.554492] env[61868]: DEBUG nova.compute.manager [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 991.554492] env[61868]: DEBUG nova.network.neutron [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 991.612482] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75393216-a1c3-4a38-b1b5-75c9e5f6e958 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.636120] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Updating instance '5d2b56d5-a4e0-4dba-9158-71036d60a40b' progress to 0 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 991.643458] env[61868]: DEBUG nova.network.neutron [req-5c16cf77-c7cc-4bf3-9654-01fd884f443e req-b77ce442-c11f-488a-b33e-a801657fed92 service nova] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Updated VIF entry in instance network info cache for port 80d0494a-2f37-48c5-b7c6-9491f5da2bc4. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 991.643458] env[61868]: DEBUG nova.network.neutron [req-5c16cf77-c7cc-4bf3-9654-01fd884f443e req-b77ce442-c11f-488a-b33e-a801657fed92 service nova] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Updating instance_info_cache with network_info: [{"id": "80d0494a-2f37-48c5-b7c6-9491f5da2bc4", "address": "fa:16:3e:25:77:13", "network": {"id": "77f33b52-e4eb-4b72-85d4-5991a2b5c36d", "bridge": "br-int", "label": "tempest-ServersTestJSON-748635995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b520c3ac58074e8d9b0bfafb817244a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap80d0494a-2f", "ovs_interfaceid": "80d0494a-2f37-48c5-b7c6-9491f5da2bc4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.678688] env[61868]: DEBUG oslo_vmware.api [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52f61d99-f642-7817-199f-718df4f164ae, 'name': SearchDatastore_Task, 'duration_secs': 0.011921} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.679029] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.679297] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 991.679559] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.682018] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.682018] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 991.682018] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2b6d6250-e81a-4d37-baac-8f987aaff4a8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.688689] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 991.688898] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 991.689685] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f1b83a36-a031-414d-aead-23f894de464b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.695621] env[61868]: DEBUG oslo_vmware.api [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 991.695621] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52751601-2859-96c5-b9e1-08806fe747d2" [ 991.695621] env[61868]: _type = "Task" [ 991.695621] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.707734] env[61868]: DEBUG oslo_vmware.api [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52751601-2859-96c5-b9e1-08806fe747d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.737427] env[61868]: DEBUG oslo_vmware.rw_handles [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52795126-451e-14d3-fb9f-527bc0f8b200/disk-0.vmdk. {{(pid=61868) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 991.737689] env[61868]: INFO nova.virt.vmwareapi.images [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Downloaded image file data 76af0b0a-47a5-487a-9b10-d16668b6caae [ 991.738524] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5428eec-ea39-48e7-aae7-5853a974c96a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.754635] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-94c58852-6832-40a5-9e6a-d429a8f5f3cf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.779577] env[61868]: INFO nova.virt.vmwareapi.images [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] The imported VM was unregistered [ 991.782150] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Caching image {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 991.783100] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Creating directory with path [datastore1] devstack-image-cache_base/76af0b0a-47a5-487a-9b10-d16668b6caae {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 991.783100] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-073e9c33-986d-40e2-b7e6-fbff0891bba6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.800552] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Created directory with path [datastore1] devstack-image-cache_base/76af0b0a-47a5-487a-9b10-d16668b6caae {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 991.800843] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_db157afc-0490-45e2-b788-b5007ddcbe1b/OSTACK_IMG_db157afc-0490-45e2-b788-b5007ddcbe1b.vmdk to [datastore1] devstack-image-cache_base/76af0b0a-47a5-487a-9b10-d16668b6caae/76af0b0a-47a5-487a-9b10-d16668b6caae.vmdk. {{(pid=61868) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 991.801329] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-26660a8d-478a-4ba7-99db-1870b6dd9b92 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.808261] env[61868]: DEBUG oslo_vmware.api [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the task: (returnval){ [ 991.808261] env[61868]: value = "task-1315841" [ 991.808261] env[61868]: _type = "Task" [ 991.808261] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.818023] env[61868]: DEBUG oslo_vmware.api [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315841, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.888981] env[61868]: DEBUG nova.policy [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fd6ed09d33f6491b83dc72cef31bfd29', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '812e115cc1114d0e8536924c4f5556f8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 991.963611] env[61868]: DEBUG nova.compute.manager [req-9a9552ff-d8f5-422e-a7e7-ddf6670858f9 req-f789c8da-e49c-4e97-89f8-827884553c4f service nova] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Received event network-changed-8f5f51e1-8b43-4ef6-b9b8-19c6b4e89e6e {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 991.963874] env[61868]: DEBUG nova.compute.manager [req-9a9552ff-d8f5-422e-a7e7-ddf6670858f9 req-f789c8da-e49c-4e97-89f8-827884553c4f service nova] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Refreshing instance network info cache due to event network-changed-8f5f51e1-8b43-4ef6-b9b8-19c6b4e89e6e. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 991.964189] env[61868]: DEBUG oslo_concurrency.lockutils [req-9a9552ff-d8f5-422e-a7e7-ddf6670858f9 req-f789c8da-e49c-4e97-89f8-827884553c4f service nova] Acquiring lock "refresh_cache-eeeec752-7e99-471d-9425-c4ccf4f6d0a5" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.964375] env[61868]: DEBUG oslo_concurrency.lockutils [req-9a9552ff-d8f5-422e-a7e7-ddf6670858f9 req-f789c8da-e49c-4e97-89f8-827884553c4f service nova] Acquired lock "refresh_cache-eeeec752-7e99-471d-9425-c4ccf4f6d0a5" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.964567] env[61868]: DEBUG nova.network.neutron [req-9a9552ff-d8f5-422e-a7e7-ddf6670858f9 req-f789c8da-e49c-4e97-89f8-827884553c4f service nova] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Refreshing network info cache for port 8f5f51e1-8b43-4ef6-b9b8-19c6b4e89e6e {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 992.054600] env[61868]: DEBUG nova.compute.manager [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 992.144804] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 992.145329] env[61868]: DEBUG oslo_concurrency.lockutils [req-5c16cf77-c7cc-4bf3-9654-01fd884f443e req-b77ce442-c11f-488a-b33e-a801657fed92 service nova] Releasing lock "refresh_cache-56703761-3421-44fd-aa63-9db36c86e69c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.146868] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "refresh_cache-75fdbc11-1ef3-442d-bcf4-55069ba9ab62" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.146868] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquired lock "refresh_cache-75fdbc11-1ef3-442d-bcf4-55069ba9ab62" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.147067] env[61868]: DEBUG nova.network.neutron [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Forcefully refreshing network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 992.148152] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ea46f4fe-5dc4-4c03-9f34-e5c306decdc5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.157778] env[61868]: DEBUG oslo_vmware.api [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 992.157778] env[61868]: value = "task-1315842" [ 992.157778] env[61868]: _type = "Task" [ 992.157778] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.168714] env[61868]: DEBUG oslo_vmware.api [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315842, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.213015] env[61868]: DEBUG oslo_vmware.api [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52751601-2859-96c5-b9e1-08806fe747d2, 'name': SearchDatastore_Task, 'duration_secs': 0.014293} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.217862] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1425eaf6-4c13-4ab9-b6f2-39d7075a1c10 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.221495] env[61868]: DEBUG nova.network.neutron [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Successfully created port: cbd81a07-4122-41bd-bb91-d47b6bc3918d {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 992.227113] env[61868]: DEBUG oslo_vmware.api [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 992.227113] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52728ad5-2dcc-62b9-4087-26cf91945f4a" [ 992.227113] env[61868]: _type = "Task" [ 992.227113] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.240405] env[61868]: DEBUG oslo_vmware.api [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52728ad5-2dcc-62b9-4087-26cf91945f4a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.272836] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfe30512-18f3-4ca3-a62b-3781a41f3ba4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.280844] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a82a713-43db-447f-ac10-09a57802f1d1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.316656] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad14547f-e98d-47e3-9565-fb3cf6e241b1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.325293] env[61868]: DEBUG oslo_vmware.api [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315841, 'name': MoveVirtualDisk_Task} progress is 12%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.329019] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-097dd60e-3245-47d6-8063-12efb5899ad0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.345238] env[61868]: DEBUG nova.compute.provider_tree [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 992.674498] env[61868]: DEBUG oslo_vmware.api [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315842, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.741837] env[61868]: DEBUG oslo_vmware.api [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52728ad5-2dcc-62b9-4087-26cf91945f4a, 'name': SearchDatastore_Task, 'duration_secs': 0.016679} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.742077] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.742416] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 56703761-3421-44fd-aa63-9db36c86e69c/56703761-3421-44fd-aa63-9db36c86e69c.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 992.742658] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-79bce575-102d-4e0f-855f-682688784b4c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.754971] env[61868]: DEBUG oslo_vmware.api [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 992.754971] env[61868]: value = "task-1315843" [ 992.754971] env[61868]: _type = "Task" [ 992.754971] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.764214] env[61868]: DEBUG nova.network.neutron [req-9a9552ff-d8f5-422e-a7e7-ddf6670858f9 req-f789c8da-e49c-4e97-89f8-827884553c4f service nova] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Updated VIF entry in instance network info cache for port 8f5f51e1-8b43-4ef6-b9b8-19c6b4e89e6e. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 992.764599] env[61868]: DEBUG nova.network.neutron [req-9a9552ff-d8f5-422e-a7e7-ddf6670858f9 req-f789c8da-e49c-4e97-89f8-827884553c4f service nova] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Updating instance_info_cache with network_info: [{"id": "8f5f51e1-8b43-4ef6-b9b8-19c6b4e89e6e", "address": "fa:16:3e:88:53:14", "network": {"id": "3535e048-fe57-4cbc-9489-0b77484069bd", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-748518849-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.202", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c8440ce785c344d793dd513b5f008d65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f5f51e1-8b", "ovs_interfaceid": "8f5f51e1-8b43-4ef6-b9b8-19c6b4e89e6e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.768787] env[61868]: DEBUG oslo_vmware.api [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315843, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.824221] env[61868]: DEBUG oslo_vmware.api [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315841, 'name': MoveVirtualDisk_Task} progress is 29%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.849895] env[61868]: DEBUG nova.scheduler.client.report [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 993.070646] env[61868]: DEBUG nova.compute.manager [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 993.097713] env[61868]: DEBUG nova.virt.hardware [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 993.098733] env[61868]: DEBUG nova.virt.hardware [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 993.099086] env[61868]: DEBUG nova.virt.hardware [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 993.099412] env[61868]: DEBUG nova.virt.hardware [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 993.099664] env[61868]: DEBUG nova.virt.hardware [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 993.100477] env[61868]: DEBUG nova.virt.hardware [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 993.100871] env[61868]: DEBUG nova.virt.hardware [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 993.101161] env[61868]: DEBUG nova.virt.hardware [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 993.101414] env[61868]: DEBUG nova.virt.hardware [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 993.101676] env[61868]: DEBUG nova.virt.hardware [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 993.101939] env[61868]: DEBUG nova.virt.hardware [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 993.102999] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ef606fe-bb57-4eeb-b7c3-80def0aeb960 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.114764] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd12546d-6542-4f71-a8a2-52d442514d07 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.176883] env[61868]: DEBUG oslo_vmware.api [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315842, 'name': PowerOffVM_Task, 'duration_secs': 0.748813} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.177418] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 993.177873] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Updating instance '5d2b56d5-a4e0-4dba-9158-71036d60a40b' progress to 17 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 993.268294] env[61868]: DEBUG oslo_vmware.api [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315843, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.270285] env[61868]: DEBUG oslo_concurrency.lockutils [req-9a9552ff-d8f5-422e-a7e7-ddf6670858f9 req-f789c8da-e49c-4e97-89f8-827884553c4f service nova] Releasing lock "refresh_cache-eeeec752-7e99-471d-9425-c4ccf4f6d0a5" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.326458] env[61868]: DEBUG oslo_vmware.api [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315841, 'name': MoveVirtualDisk_Task} progress is 52%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.355688] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.313s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.356775] env[61868]: DEBUG nova.compute.manager [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 993.423999] env[61868]: DEBUG nova.network.neutron [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Updating instance_info_cache with network_info: [{"id": "21d6cab8-f691-425c-bfa8-9d2f1c294b64", "address": "fa:16:3e:76:a4:70", "network": {"id": "e7457f1a-171a-4c82-a260-1d1a7a95861e", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1032778554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b9b800086764b41a8817d1d8e463a2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21d6cab8-f6", "ovs_interfaceid": "21d6cab8-f691-425c-bfa8-9d2f1c294b64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.685110] env[61868]: DEBUG nova.virt.hardware [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 993.685378] env[61868]: DEBUG nova.virt.hardware [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 993.685586] env[61868]: DEBUG nova.virt.hardware [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 993.685726] env[61868]: DEBUG nova.virt.hardware [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 993.685876] env[61868]: DEBUG nova.virt.hardware [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 993.686042] env[61868]: DEBUG nova.virt.hardware [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 993.686260] env[61868]: DEBUG nova.virt.hardware [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 993.686439] env[61868]: DEBUG nova.virt.hardware [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 993.686609] env[61868]: DEBUG nova.virt.hardware [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 993.686774] env[61868]: DEBUG nova.virt.hardware [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 993.686948] env[61868]: DEBUG nova.virt.hardware [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 993.692487] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-162470e2-cc44-4f31-b881-2e7ef4ae6041 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.711231] env[61868]: DEBUG oslo_vmware.api [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 993.711231] env[61868]: value = "task-1315844" [ 993.711231] env[61868]: _type = "Task" [ 993.711231] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.723816] env[61868]: DEBUG oslo_vmware.api [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315844, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.766889] env[61868]: DEBUG oslo_vmware.api [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315843, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.71473} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.767222] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 56703761-3421-44fd-aa63-9db36c86e69c/56703761-3421-44fd-aa63-9db36c86e69c.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 993.767454] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 993.767792] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1fefbfee-7cac-4deb-8241-1e152c099f31 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.776487] env[61868]: DEBUG oslo_vmware.api [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 993.776487] env[61868]: value = "task-1315845" [ 993.776487] env[61868]: _type = "Task" [ 993.776487] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.787817] env[61868]: DEBUG oslo_vmware.api [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315845, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.827247] env[61868]: DEBUG oslo_vmware.api [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315841, 'name': MoveVirtualDisk_Task} progress is 71%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.864322] env[61868]: DEBUG nova.compute.utils [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 993.865224] env[61868]: DEBUG nova.compute.manager [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 993.865542] env[61868]: DEBUG nova.network.neutron [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 993.913169] env[61868]: DEBUG nova.policy [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c14b4ff01717495ca97c7f35f91c2995', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd7fc76299baf4a97b57139e5f1caa16e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 993.916079] env[61868]: DEBUG nova.network.neutron [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Successfully updated port: cbd81a07-4122-41bd-bb91-d47b6bc3918d {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 993.926945] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Releasing lock "refresh_cache-75fdbc11-1ef3-442d-bcf4-55069ba9ab62" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.927164] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Updated the network info_cache for instance {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 993.927364] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 993.927899] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 993.928130] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 993.928329] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 993.928522] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 993.928708] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 993.928951] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61868) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 993.929191] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager.update_available_resource {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 993.990492] env[61868]: DEBUG nova.compute.manager [req-e5b0171a-43d7-46b9-bc4b-8b50ff211823 req-fd5d2e0d-c643-418a-b34f-913665271e05 service nova] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Received event network-vif-plugged-cbd81a07-4122-41bd-bb91-d47b6bc3918d {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 993.990721] env[61868]: DEBUG oslo_concurrency.lockutils [req-e5b0171a-43d7-46b9-bc4b-8b50ff211823 req-fd5d2e0d-c643-418a-b34f-913665271e05 service nova] Acquiring lock "54dacd73-c760-49bd-9baf-ad2012a53264-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.990937] env[61868]: DEBUG oslo_concurrency.lockutils [req-e5b0171a-43d7-46b9-bc4b-8b50ff211823 req-fd5d2e0d-c643-418a-b34f-913665271e05 service nova] Lock "54dacd73-c760-49bd-9baf-ad2012a53264-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.991134] env[61868]: DEBUG oslo_concurrency.lockutils [req-e5b0171a-43d7-46b9-bc4b-8b50ff211823 req-fd5d2e0d-c643-418a-b34f-913665271e05 service nova] Lock "54dacd73-c760-49bd-9baf-ad2012a53264-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.991299] env[61868]: DEBUG nova.compute.manager [req-e5b0171a-43d7-46b9-bc4b-8b50ff211823 req-fd5d2e0d-c643-418a-b34f-913665271e05 service nova] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] No waiting events found dispatching network-vif-plugged-cbd81a07-4122-41bd-bb91-d47b6bc3918d {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 993.991474] env[61868]: WARNING nova.compute.manager [req-e5b0171a-43d7-46b9-bc4b-8b50ff211823 req-fd5d2e0d-c643-418a-b34f-913665271e05 service nova] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Received unexpected event network-vif-plugged-cbd81a07-4122-41bd-bb91-d47b6bc3918d for instance with vm_state building and task_state spawning. [ 993.991654] env[61868]: DEBUG nova.compute.manager [req-e5b0171a-43d7-46b9-bc4b-8b50ff211823 req-fd5d2e0d-c643-418a-b34f-913665271e05 service nova] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Received event network-changed-cbd81a07-4122-41bd-bb91-d47b6bc3918d {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 993.991779] env[61868]: DEBUG nova.compute.manager [req-e5b0171a-43d7-46b9-bc4b-8b50ff211823 req-fd5d2e0d-c643-418a-b34f-913665271e05 service nova] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Refreshing instance network info cache due to event network-changed-cbd81a07-4122-41bd-bb91-d47b6bc3918d. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 993.991984] env[61868]: DEBUG oslo_concurrency.lockutils [req-e5b0171a-43d7-46b9-bc4b-8b50ff211823 req-fd5d2e0d-c643-418a-b34f-913665271e05 service nova] Acquiring lock "refresh_cache-54dacd73-c760-49bd-9baf-ad2012a53264" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.992141] env[61868]: DEBUG oslo_concurrency.lockutils [req-e5b0171a-43d7-46b9-bc4b-8b50ff211823 req-fd5d2e0d-c643-418a-b34f-913665271e05 service nova] Acquired lock "refresh_cache-54dacd73-c760-49bd-9baf-ad2012a53264" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.992362] env[61868]: DEBUG nova.network.neutron [req-e5b0171a-43d7-46b9-bc4b-8b50ff211823 req-fd5d2e0d-c643-418a-b34f-913665271e05 service nova] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Refreshing network info cache for port cbd81a07-4122-41bd-bb91-d47b6bc3918d {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 994.227562] env[61868]: DEBUG oslo_vmware.api [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315844, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.230062] env[61868]: DEBUG nova.network.neutron [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Successfully created port: f8f4c521-03cd-4e54-99f1-761d9a515aa3 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 994.289029] env[61868]: DEBUG oslo_vmware.api [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315845, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069799} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.289500] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 994.290466] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad36c77b-c867-46a7-b8d7-519c172b48ee {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.315881] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] 56703761-3421-44fd-aa63-9db36c86e69c/56703761-3421-44fd-aa63-9db36c86e69c.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 994.316789] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-32f01efd-5c8a-4f7e-80cc-c76088abed2f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.342982] env[61868]: DEBUG oslo_vmware.api [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315841, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.345344] env[61868]: DEBUG oslo_vmware.api [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 994.345344] env[61868]: value = "task-1315846" [ 994.345344] env[61868]: _type = "Task" [ 994.345344] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.357096] env[61868]: DEBUG oslo_vmware.api [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315846, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.366445] env[61868]: DEBUG nova.compute.manager [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 994.419222] env[61868]: DEBUG oslo_concurrency.lockutils [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "refresh_cache-54dacd73-c760-49bd-9baf-ad2012a53264" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.433056] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.433056] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.433056] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.433056] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61868) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 994.434718] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cf4f7ad-e3e5-4ff2-a1b1-fccbe9cf9acc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.444719] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d191831-3045-453d-ac29-08b70d1a417e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.462916] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfe02abc-8a04-48c9-a2b4-5459dae388cf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.475429] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddbd8618-6fce-4793-af94-3b6980e7c7f1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.516645] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181238MB free_disk=134GB free_vcpus=48 pci_devices=None {{(pid=61868) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 994.517129] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.517505] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.569758] env[61868]: DEBUG nova.network.neutron [req-e5b0171a-43d7-46b9-bc4b-8b50ff211823 req-fd5d2e0d-c643-418a-b34f-913665271e05 service nova] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 994.700182] env[61868]: DEBUG nova.network.neutron [req-e5b0171a-43d7-46b9-bc4b-8b50ff211823 req-fd5d2e0d-c643-418a-b34f-913665271e05 service nova] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.722993] env[61868]: DEBUG oslo_vmware.api [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315844, 'name': ReconfigVM_Task, 'duration_secs': 0.798052} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.723629] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Updating instance '5d2b56d5-a4e0-4dba-9158-71036d60a40b' progress to 33 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 994.826538] env[61868]: DEBUG oslo_vmware.api [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315841, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.858404} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.826827] env[61868]: INFO nova.virt.vmwareapi.ds_util [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_db157afc-0490-45e2-b788-b5007ddcbe1b/OSTACK_IMG_db157afc-0490-45e2-b788-b5007ddcbe1b.vmdk to [datastore1] devstack-image-cache_base/76af0b0a-47a5-487a-9b10-d16668b6caae/76af0b0a-47a5-487a-9b10-d16668b6caae.vmdk. [ 994.827025] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Cleaning up location [datastore1] OSTACK_IMG_db157afc-0490-45e2-b788-b5007ddcbe1b {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 994.827194] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_db157afc-0490-45e2-b788-b5007ddcbe1b {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 994.827455] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ab49b9ac-d579-4526-a30f-4f592f52b861 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.834143] env[61868]: DEBUG oslo_vmware.api [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the task: (returnval){ [ 994.834143] env[61868]: value = "task-1315847" [ 994.834143] env[61868]: _type = "Task" [ 994.834143] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.841938] env[61868]: DEBUG oslo_vmware.api [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315847, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.854443] env[61868]: DEBUG oslo_vmware.api [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315846, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.204722] env[61868]: DEBUG oslo_concurrency.lockutils [req-e5b0171a-43d7-46b9-bc4b-8b50ff211823 req-fd5d2e0d-c643-418a-b34f-913665271e05 service nova] Releasing lock "refresh_cache-54dacd73-c760-49bd-9baf-ad2012a53264" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.205153] env[61868]: DEBUG oslo_concurrency.lockutils [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquired lock "refresh_cache-54dacd73-c760-49bd-9baf-ad2012a53264" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.205322] env[61868]: DEBUG nova.network.neutron [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 995.231026] env[61868]: DEBUG nova.virt.hardware [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 995.231268] env[61868]: DEBUG nova.virt.hardware [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 995.231323] env[61868]: DEBUG nova.virt.hardware [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 995.231630] env[61868]: DEBUG nova.virt.hardware [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 995.231834] env[61868]: DEBUG nova.virt.hardware [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 995.231968] env[61868]: DEBUG nova.virt.hardware [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 995.232249] env[61868]: DEBUG nova.virt.hardware [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 995.232495] env[61868]: DEBUG nova.virt.hardware [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 995.232690] env[61868]: DEBUG nova.virt.hardware [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 995.232864] env[61868]: DEBUG nova.virt.hardware [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 995.233116] env[61868]: DEBUG nova.virt.hardware [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 995.239396] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Reconfiguring VM instance instance-00000058 to detach disk 2000 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 995.240098] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7ef93a6e-d0a6-41c4-8798-735c617ad7e7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.262694] env[61868]: DEBUG oslo_vmware.api [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 995.262694] env[61868]: value = "task-1315848" [ 995.262694] env[61868]: _type = "Task" [ 995.262694] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.274508] env[61868]: DEBUG oslo_vmware.api [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315848, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.343212] env[61868]: DEBUG oslo_vmware.api [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315847, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.055007} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.343577] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 995.343715] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Releasing lock "[datastore1] devstack-image-cache_base/76af0b0a-47a5-487a-9b10-d16668b6caae/76af0b0a-47a5-487a-9b10-d16668b6caae.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.343970] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/76af0b0a-47a5-487a-9b10-d16668b6caae/76af0b0a-47a5-487a-9b10-d16668b6caae.vmdk to [datastore1] 75fdbc11-1ef3-442d-bcf4-55069ba9ab62/75fdbc11-1ef3-442d-bcf4-55069ba9ab62.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 995.344246] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b5bcf827-3272-4dfb-a08b-804f97d1fe9b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.354357] env[61868]: DEBUG oslo_vmware.api [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315846, 'name': ReconfigVM_Task, 'duration_secs': 0.580619} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.355526] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Reconfigured VM instance instance-0000005d to attach disk [datastore2] 56703761-3421-44fd-aa63-9db36c86e69c/56703761-3421-44fd-aa63-9db36c86e69c.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 995.356230] env[61868]: DEBUG oslo_vmware.api [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the task: (returnval){ [ 995.356230] env[61868]: value = "task-1315849" [ 995.356230] env[61868]: _type = "Task" [ 995.356230] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.356432] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c0d951dc-4035-421d-ae75-4c60d49aa054 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.366064] env[61868]: DEBUG oslo_vmware.api [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315849, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.367328] env[61868]: DEBUG oslo_vmware.api [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 995.367328] env[61868]: value = "task-1315850" [ 995.367328] env[61868]: _type = "Task" [ 995.367328] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.375251] env[61868]: DEBUG oslo_vmware.api [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315850, 'name': Rename_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.378579] env[61868]: DEBUG nova.compute.manager [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 995.399970] env[61868]: DEBUG nova.virt.hardware [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 995.400254] env[61868]: DEBUG nova.virt.hardware [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 995.400417] env[61868]: DEBUG nova.virt.hardware [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 995.400646] env[61868]: DEBUG nova.virt.hardware [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 995.400824] env[61868]: DEBUG nova.virt.hardware [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 995.400971] env[61868]: DEBUG nova.virt.hardware [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 995.401198] env[61868]: DEBUG nova.virt.hardware [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 995.401361] env[61868]: DEBUG nova.virt.hardware [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 995.401536] env[61868]: DEBUG nova.virt.hardware [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 995.401701] env[61868]: DEBUG nova.virt.hardware [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 995.401878] env[61868]: DEBUG nova.virt.hardware [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 995.402821] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7391e94-558c-4c23-922b-26ed8658cfcd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.411134] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d539972-31a5-4eb0-a390-eecd97d06874 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.529380] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Applying migration context for instance 5d2b56d5-a4e0-4dba-9158-71036d60a40b as it has an incoming, in-progress migration 9f4d2d6e-1376-4d88-8eb6-b575d737488f. Migration status is migrating {{(pid=61868) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 995.531710] env[61868]: INFO nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Updating resource usage from migration 9f4d2d6e-1376-4d88-8eb6-b575d737488f [ 995.556614] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance e59a182d-97b2-454a-bc40-8afb0839324f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 995.556789] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 9286cc76-1152-4fd4-9279-33d6ee4b9486 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 995.556918] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance a96cab5a-c053-4e45-96f6-2aba0a819110 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 995.557050] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance cbeeb7b1-166b-49f7-88c5-6ed05c9dc759 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 995.557171] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 75fdbc11-1ef3-442d-bcf4-55069ba9ab62 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 995.557289] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance eeeec752-7e99-471d-9425-c4ccf4f6d0a5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 995.557405] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 56703761-3421-44fd-aa63-9db36c86e69c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 995.557517] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Migration 9f4d2d6e-1376-4d88-8eb6-b575d737488f is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 995.557627] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 5d2b56d5-a4e0-4dba-9158-71036d60a40b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 995.557788] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 54dacd73-c760-49bd-9baf-ad2012a53264 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 995.557978] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 995.558216] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Total usable vcpus: 48, total allocated vcpus: 11 {{(pid=61868) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 995.558358] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2688MB phys_disk=200GB used_disk=11GB total_vcpus=48 used_vcpus=11 pci_stats=[] {{(pid=61868) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 995.711539] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-082ce641-520f-42ee-bfcc-fffba937cf6d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.722173] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e286c71c-00ca-4381-bf18-1e9a6cbc1e06 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.755687] env[61868]: DEBUG nova.network.neutron [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 995.758219] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b22aaff-6e71-449d-a0b3-b37b966016d0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.772686] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8a8c99c-0d0b-471c-83ca-838ba15a3219 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.780156] env[61868]: DEBUG oslo_vmware.api [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315848, 'name': ReconfigVM_Task, 'duration_secs': 0.235409} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.782967] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Reconfigured VM instance instance-00000058 to detach disk 2000 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 995.784386] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-063dd077-272d-4ea1-bba0-a9ab0da0c0a6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.797379] env[61868]: DEBUG nova.compute.provider_tree [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 995.800144] env[61868]: DEBUG nova.compute.manager [req-04851bf2-942d-4677-9541-ff80c01592f7 req-bbe79422-8a0a-46c6-a161-6352c8ece501 service nova] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Received event network-vif-plugged-f8f4c521-03cd-4e54-99f1-761d9a515aa3 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 995.801172] env[61868]: DEBUG oslo_concurrency.lockutils [req-04851bf2-942d-4677-9541-ff80c01592f7 req-bbe79422-8a0a-46c6-a161-6352c8ece501 service nova] Acquiring lock "69a0b94b-2bf4-49ad-8fb0-ea945176e5eb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.801172] env[61868]: DEBUG oslo_concurrency.lockutils [req-04851bf2-942d-4677-9541-ff80c01592f7 req-bbe79422-8a0a-46c6-a161-6352c8ece501 service nova] Lock "69a0b94b-2bf4-49ad-8fb0-ea945176e5eb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.801172] env[61868]: DEBUG oslo_concurrency.lockutils [req-04851bf2-942d-4677-9541-ff80c01592f7 req-bbe79422-8a0a-46c6-a161-6352c8ece501 service nova] Lock "69a0b94b-2bf4-49ad-8fb0-ea945176e5eb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.802060] env[61868]: DEBUG nova.compute.manager [req-04851bf2-942d-4677-9541-ff80c01592f7 req-bbe79422-8a0a-46c6-a161-6352c8ece501 service nova] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] No waiting events found dispatching network-vif-plugged-f8f4c521-03cd-4e54-99f1-761d9a515aa3 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 995.802060] env[61868]: WARNING nova.compute.manager [req-04851bf2-942d-4677-9541-ff80c01592f7 req-bbe79422-8a0a-46c6-a161-6352c8ece501 service nova] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Received unexpected event network-vif-plugged-f8f4c521-03cd-4e54-99f1-761d9a515aa3 for instance with vm_state building and task_state spawning. [ 995.823853] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] 5d2b56d5-a4e0-4dba-9158-71036d60a40b/5d2b56d5-a4e0-4dba-9158-71036d60a40b.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 995.825343] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4b81e2b7-319d-4d50-ad66-573594cf5784 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.848283] env[61868]: DEBUG oslo_vmware.api [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 995.848283] env[61868]: value = "task-1315851" [ 995.848283] env[61868]: _type = "Task" [ 995.848283] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.857520] env[61868]: DEBUG oslo_vmware.api [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315851, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.865980] env[61868]: DEBUG oslo_vmware.api [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315849, 'name': CopyVirtualDisk_Task} progress is 15%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.876570] env[61868]: DEBUG oslo_vmware.api [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315850, 'name': Rename_Task, 'duration_secs': 0.211498} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.876856] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 995.877149] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-17f02466-94c5-4095-a4b4-8a1ef0455a40 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.884224] env[61868]: DEBUG oslo_vmware.api [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 995.884224] env[61868]: value = "task-1315852" [ 995.884224] env[61868]: _type = "Task" [ 995.884224] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.891337] env[61868]: DEBUG nova.network.neutron [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Successfully updated port: f8f4c521-03cd-4e54-99f1-761d9a515aa3 {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 995.898268] env[61868]: DEBUG oslo_vmware.api [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315852, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.977788] env[61868]: DEBUG nova.network.neutron [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Updating instance_info_cache with network_info: [{"id": "cbd81a07-4122-41bd-bb91-d47b6bc3918d", "address": "fa:16:3e:31:4b:86", "network": {"id": "5b4c5906-b873-4369-855b-e0898ad84812", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-254512637-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "812e115cc1114d0e8536924c4f5556f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcbd81a07-41", "ovs_interfaceid": "cbd81a07-4122-41bd-bb91-d47b6bc3918d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.300834] env[61868]: DEBUG nova.scheduler.client.report [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 996.359023] env[61868]: DEBUG oslo_vmware.api [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315851, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.367090] env[61868]: DEBUG oslo_vmware.api [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315849, 'name': CopyVirtualDisk_Task} progress is 38%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.393928] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "refresh_cache-69a0b94b-2bf4-49ad-8fb0-ea945176e5eb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.394141] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquired lock "refresh_cache-69a0b94b-2bf4-49ad-8fb0-ea945176e5eb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.394292] env[61868]: DEBUG nova.network.neutron [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 996.396087] env[61868]: DEBUG oslo_vmware.api [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315852, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.481059] env[61868]: DEBUG oslo_concurrency.lockutils [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Releasing lock "refresh_cache-54dacd73-c760-49bd-9baf-ad2012a53264" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.481631] env[61868]: DEBUG nova.compute.manager [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Instance network_info: |[{"id": "cbd81a07-4122-41bd-bb91-d47b6bc3918d", "address": "fa:16:3e:31:4b:86", "network": {"id": "5b4c5906-b873-4369-855b-e0898ad84812", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-254512637-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "812e115cc1114d0e8536924c4f5556f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcbd81a07-41", "ovs_interfaceid": "cbd81a07-4122-41bd-bb91-d47b6bc3918d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 996.482369] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:31:4b:86', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a92a4ffe-7939-4697-bf98-5b22e2c7feda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cbd81a07-4122-41bd-bb91-d47b6bc3918d', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 996.495621] env[61868]: DEBUG oslo.service.loopingcall [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 996.496120] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 996.496574] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-da66567f-b6a6-46d0-b8e9-082dadbfd326 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.528335] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 996.528335] env[61868]: value = "task-1315853" [ 996.528335] env[61868]: _type = "Task" [ 996.528335] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.540761] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315853, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.806043] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61868) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 996.806250] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.289s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.860558] env[61868]: DEBUG oslo_vmware.api [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315851, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.868722] env[61868]: DEBUG oslo_vmware.api [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315849, 'name': CopyVirtualDisk_Task} progress is 57%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.894287] env[61868]: DEBUG oslo_vmware.api [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315852, 'name': PowerOnVM_Task, 'duration_secs': 0.644565} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.894593] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 996.894848] env[61868]: INFO nova.compute.manager [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Took 8.40 seconds to spawn the instance on the hypervisor. [ 996.895054] env[61868]: DEBUG nova.compute.manager [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 996.895904] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea37e062-d6e3-49de-bc1e-9ab59209820d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.937501] env[61868]: DEBUG nova.network.neutron [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 997.039309] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315853, 'name': CreateVM_Task, 'duration_secs': 0.461557} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.039551] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 997.040310] env[61868]: DEBUG oslo_concurrency.lockutils [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.040465] env[61868]: DEBUG oslo_concurrency.lockutils [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.040799] env[61868]: DEBUG oslo_concurrency.lockutils [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 997.041078] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5bf2ce30-3abb-4fe6-b703-929cf821b001 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.046233] env[61868]: DEBUG oslo_vmware.api [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 997.046233] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]529f6ed2-26a8-16ab-0449-47124b9d78fa" [ 997.046233] env[61868]: _type = "Task" [ 997.046233] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.055930] env[61868]: DEBUG oslo_vmware.api [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]529f6ed2-26a8-16ab-0449-47124b9d78fa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.164306] env[61868]: DEBUG nova.network.neutron [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Updating instance_info_cache with network_info: [{"id": "f8f4c521-03cd-4e54-99f1-761d9a515aa3", "address": "fa:16:3e:3c:dd:65", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8f4c521-03", "ovs_interfaceid": "f8f4c521-03cd-4e54-99f1-761d9a515aa3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.360082] env[61868]: DEBUG oslo_vmware.api [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315851, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.368314] env[61868]: DEBUG oslo_vmware.api [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315849, 'name': CopyVirtualDisk_Task} progress is 80%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.416507] env[61868]: INFO nova.compute.manager [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Took 18.67 seconds to build instance. [ 997.557772] env[61868]: DEBUG oslo_vmware.api [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]529f6ed2-26a8-16ab-0449-47124b9d78fa, 'name': SearchDatastore_Task, 'duration_secs': 0.015608} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.558188] env[61868]: DEBUG oslo_concurrency.lockutils [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.558448] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 997.558695] env[61868]: DEBUG oslo_concurrency.lockutils [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.558846] env[61868]: DEBUG oslo_concurrency.lockutils [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.559044] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 997.559332] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4faf3a0a-0de6-4935-9dcb-ef7d612e9768 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.568331] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 997.568558] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 997.569308] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de4a17c4-0b60-46dd-99db-a5542e1d9c55 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.575150] env[61868]: DEBUG oslo_vmware.api [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 997.575150] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]526a0711-fa15-044d-e4be-34b58269eab0" [ 997.575150] env[61868]: _type = "Task" [ 997.575150] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.583233] env[61868]: DEBUG oslo_vmware.api [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]526a0711-fa15-044d-e4be-34b58269eab0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.668453] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Releasing lock "refresh_cache-69a0b94b-2bf4-49ad-8fb0-ea945176e5eb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.668811] env[61868]: DEBUG nova.compute.manager [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Instance network_info: |[{"id": "f8f4c521-03cd-4e54-99f1-761d9a515aa3", "address": "fa:16:3e:3c:dd:65", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8f4c521-03", "ovs_interfaceid": "f8f4c521-03cd-4e54-99f1-761d9a515aa3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 997.669288] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3c:dd:65', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8b29df12-5674-476d-a9e5-5e20f704d224', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f8f4c521-03cd-4e54-99f1-761d9a515aa3', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 997.677117] env[61868]: DEBUG oslo.service.loopingcall [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 997.677396] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 997.677638] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e7cf121b-b6ac-4b92-b8eb-a866eab0299b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.697767] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 997.697767] env[61868]: value = "task-1315854" [ 997.697767] env[61868]: _type = "Task" [ 997.697767] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.707490] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315854, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.814858] env[61868]: DEBUG nova.compute.manager [req-4cfcd72e-1432-403c-acc3-2a6c46069b08 req-b0f45cba-ea5a-4217-8c3f-d8bdd0ba9952 service nova] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Received event network-changed-f8f4c521-03cd-4e54-99f1-761d9a515aa3 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 997.815098] env[61868]: DEBUG nova.compute.manager [req-4cfcd72e-1432-403c-acc3-2a6c46069b08 req-b0f45cba-ea5a-4217-8c3f-d8bdd0ba9952 service nova] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Refreshing instance network info cache due to event network-changed-f8f4c521-03cd-4e54-99f1-761d9a515aa3. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 997.815391] env[61868]: DEBUG oslo_concurrency.lockutils [req-4cfcd72e-1432-403c-acc3-2a6c46069b08 req-b0f45cba-ea5a-4217-8c3f-d8bdd0ba9952 service nova] Acquiring lock "refresh_cache-69a0b94b-2bf4-49ad-8fb0-ea945176e5eb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.815580] env[61868]: DEBUG oslo_concurrency.lockutils [req-4cfcd72e-1432-403c-acc3-2a6c46069b08 req-b0f45cba-ea5a-4217-8c3f-d8bdd0ba9952 service nova] Acquired lock "refresh_cache-69a0b94b-2bf4-49ad-8fb0-ea945176e5eb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.816116] env[61868]: DEBUG nova.network.neutron [req-4cfcd72e-1432-403c-acc3-2a6c46069b08 req-b0f45cba-ea5a-4217-8c3f-d8bdd0ba9952 service nova] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Refreshing network info cache for port f8f4c521-03cd-4e54-99f1-761d9a515aa3 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 997.860381] env[61868]: DEBUG oslo_vmware.api [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315851, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.867875] env[61868]: DEBUG oslo_vmware.api [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315849, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.481405} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.868154] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/76af0b0a-47a5-487a-9b10-d16668b6caae/76af0b0a-47a5-487a-9b10-d16668b6caae.vmdk to [datastore1] 75fdbc11-1ef3-442d-bcf4-55069ba9ab62/75fdbc11-1ef3-442d-bcf4-55069ba9ab62.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 997.868920] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b618c3d7-2fb1-4f68-b237-9f6fbb0dd788 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.890896] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 75fdbc11-1ef3-442d-bcf4-55069ba9ab62/75fdbc11-1ef3-442d-bcf4-55069ba9ab62.vmdk or device None with type streamOptimized {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 997.891590] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4fd279e7-599d-4ef8-bf70-8d10b3fda0c4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.910152] env[61868]: DEBUG oslo_vmware.api [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the task: (returnval){ [ 997.910152] env[61868]: value = "task-1315855" [ 997.910152] env[61868]: _type = "Task" [ 997.910152] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.919271] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d2909ba3-1298-47ed-b073-4c4fda443727 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "56703761-3421-44fd-aa63-9db36c86e69c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.192s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.919572] env[61868]: DEBUG oslo_vmware.api [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315855, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.086275] env[61868]: DEBUG oslo_vmware.api [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]526a0711-fa15-044d-e4be-34b58269eab0, 'name': SearchDatastore_Task, 'duration_secs': 0.013676} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.087154] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd8b2ab4-e7aa-47e3-adf4-f0e3b2777f0c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.092429] env[61868]: DEBUG oslo_vmware.api [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 998.092429] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]527e87cc-c024-5038-1dd6-054914c3b7de" [ 998.092429] env[61868]: _type = "Task" [ 998.092429] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.101038] env[61868]: DEBUG oslo_vmware.api [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]527e87cc-c024-5038-1dd6-054914c3b7de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.207453] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315854, 'name': CreateVM_Task, 'duration_secs': 0.421269} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.207669] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 998.209468] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 998.209675] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.210078] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 998.210435] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0616996a-9660-4d6f-80d9-f95df072c6fb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.215281] env[61868]: DEBUG oslo_vmware.api [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for the task: (returnval){ [ 998.215281] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52361c9f-5814-f973-a019-6f6d423c9fd8" [ 998.215281] env[61868]: _type = "Task" [ 998.215281] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.223959] env[61868]: DEBUG oslo_vmware.api [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52361c9f-5814-f973-a019-6f6d423c9fd8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.361085] env[61868]: DEBUG oslo_vmware.api [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315851, 'name': ReconfigVM_Task, 'duration_secs': 2.106878} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.361460] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Reconfigured VM instance instance-00000058 to attach disk [datastore1] 5d2b56d5-a4e0-4dba-9158-71036d60a40b/5d2b56d5-a4e0-4dba-9158-71036d60a40b.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 998.361784] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Updating instance '5d2b56d5-a4e0-4dba-9158-71036d60a40b' progress to 50 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 998.423977] env[61868]: DEBUG oslo_vmware.api [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315855, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.609272] env[61868]: DEBUG oslo_vmware.api [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]527e87cc-c024-5038-1dd6-054914c3b7de, 'name': SearchDatastore_Task, 'duration_secs': 0.012964} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.613089] env[61868]: DEBUG oslo_concurrency.lockutils [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.613451] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 54dacd73-c760-49bd-9baf-ad2012a53264/54dacd73-c760-49bd-9baf-ad2012a53264.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 998.614079] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-93cae8bb-5d74-470b-8509-866cb2f4ed5f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.621643] env[61868]: DEBUG oslo_vmware.api [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 998.621643] env[61868]: value = "task-1315856" [ 998.621643] env[61868]: _type = "Task" [ 998.621643] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.630214] env[61868]: DEBUG oslo_vmware.api [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315856, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.726060] env[61868]: DEBUG oslo_vmware.api [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52361c9f-5814-f973-a019-6f6d423c9fd8, 'name': SearchDatastore_Task, 'duration_secs': 0.012531} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.726584] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.726912] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 998.727218] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 998.727417] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.727676] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 998.728029] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-01d18cfd-6925-443c-9505-771aef11c500 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.741395] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 998.741639] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 998.742476] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-863dc2e3-f336-46af-8473-7cf1b9e576fd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.747641] env[61868]: DEBUG oslo_vmware.api [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for the task: (returnval){ [ 998.747641] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]523eb7b4-e138-c8cc-3cb5-c6f336c588ba" [ 998.747641] env[61868]: _type = "Task" [ 998.747641] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.756483] env[61868]: DEBUG oslo_vmware.api [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]523eb7b4-e138-c8cc-3cb5-c6f336c588ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.786672] env[61868]: DEBUG nova.network.neutron [req-4cfcd72e-1432-403c-acc3-2a6c46069b08 req-b0f45cba-ea5a-4217-8c3f-d8bdd0ba9952 service nova] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Updated VIF entry in instance network info cache for port f8f4c521-03cd-4e54-99f1-761d9a515aa3. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 998.786672] env[61868]: DEBUG nova.network.neutron [req-4cfcd72e-1432-403c-acc3-2a6c46069b08 req-b0f45cba-ea5a-4217-8c3f-d8bdd0ba9952 service nova] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Updating instance_info_cache with network_info: [{"id": "f8f4c521-03cd-4e54-99f1-761d9a515aa3", "address": "fa:16:3e:3c:dd:65", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8f4c521-03", "ovs_interfaceid": "f8f4c521-03cd-4e54-99f1-761d9a515aa3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.869375] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac0175fc-bc49-47e5-9f41-f8d3fd8b5b0c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.891647] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff68268e-ca14-4b48-800d-0b1f4d0cb098 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.910042] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Updating instance '5d2b56d5-a4e0-4dba-9158-71036d60a40b' progress to 67 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 998.922044] env[61868]: DEBUG oslo_vmware.api [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315855, 'name': ReconfigVM_Task, 'duration_secs': 0.555832} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.922247] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 75fdbc11-1ef3-442d-bcf4-55069ba9ab62/75fdbc11-1ef3-442d-bcf4-55069ba9ab62.vmdk or device None with type streamOptimized {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 998.922886] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4181c7eb-27aa-461e-bf57-10a057f4cf8d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.931306] env[61868]: DEBUG oslo_vmware.api [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the task: (returnval){ [ 998.931306] env[61868]: value = "task-1315857" [ 998.931306] env[61868]: _type = "Task" [ 998.931306] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.940852] env[61868]: DEBUG oslo_vmware.api [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315857, 'name': Rename_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.101145] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7b1de24a-aefe-4a43-a01d-a2b914557110 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "56703761-3421-44fd-aa63-9db36c86e69c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.101509] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7b1de24a-aefe-4a43-a01d-a2b914557110 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "56703761-3421-44fd-aa63-9db36c86e69c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.101783] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7b1de24a-aefe-4a43-a01d-a2b914557110 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "56703761-3421-44fd-aa63-9db36c86e69c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.102030] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7b1de24a-aefe-4a43-a01d-a2b914557110 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "56703761-3421-44fd-aa63-9db36c86e69c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.102262] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7b1de24a-aefe-4a43-a01d-a2b914557110 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "56703761-3421-44fd-aa63-9db36c86e69c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.104927] env[61868]: INFO nova.compute.manager [None req-7b1de24a-aefe-4a43-a01d-a2b914557110 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Terminating instance [ 999.107075] env[61868]: DEBUG nova.compute.manager [None req-7b1de24a-aefe-4a43-a01d-a2b914557110 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 999.107378] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-7b1de24a-aefe-4a43-a01d-a2b914557110 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 999.108352] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f8962fb-8484-45f8-94c0-3a62b068ec45 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.117418] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b1de24a-aefe-4a43-a01d-a2b914557110 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 999.117798] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-77de0d96-2ce7-4f31-bd85-e6316c81084e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.129861] env[61868]: DEBUG oslo_vmware.api [None req-7b1de24a-aefe-4a43-a01d-a2b914557110 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 999.129861] env[61868]: value = "task-1315858" [ 999.129861] env[61868]: _type = "Task" [ 999.129861] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.137322] env[61868]: DEBUG oslo_vmware.api [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315856, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.142882] env[61868]: DEBUG oslo_vmware.api [None req-7b1de24a-aefe-4a43-a01d-a2b914557110 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315858, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.259203] env[61868]: DEBUG oslo_vmware.api [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]523eb7b4-e138-c8cc-3cb5-c6f336c588ba, 'name': SearchDatastore_Task, 'duration_secs': 0.023569} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.260100] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3fb8c7db-0a9f-44ea-871c-cc9ee7f0f4c0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.265943] env[61868]: DEBUG oslo_vmware.api [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for the task: (returnval){ [ 999.265943] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52606736-412e-1bef-434a-c3dc7740cbd7" [ 999.265943] env[61868]: _type = "Task" [ 999.265943] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.274383] env[61868]: DEBUG oslo_vmware.api [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52606736-412e-1bef-434a-c3dc7740cbd7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.288133] env[61868]: DEBUG oslo_concurrency.lockutils [req-4cfcd72e-1432-403c-acc3-2a6c46069b08 req-b0f45cba-ea5a-4217-8c3f-d8bdd0ba9952 service nova] Releasing lock "refresh_cache-69a0b94b-2bf4-49ad-8fb0-ea945176e5eb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.441013] env[61868]: DEBUG oslo_vmware.api [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315857, 'name': Rename_Task, 'duration_secs': 0.170811} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.441378] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 999.441653] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-42449ec4-1dc4-4410-95fe-9273c2637048 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.447862] env[61868]: DEBUG oslo_vmware.api [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the task: (returnval){ [ 999.447862] env[61868]: value = "task-1315859" [ 999.447862] env[61868]: _type = "Task" [ 999.447862] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.456264] env[61868]: DEBUG oslo_vmware.api [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315859, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.477831] env[61868]: DEBUG nova.network.neutron [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Port 9023a9b8-ad2a-42b6-b2cd-b2cdb86dd9cb binding to destination host cpu-1 is already ACTIVE {{(pid=61868) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 999.632255] env[61868]: DEBUG oslo_vmware.api [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315856, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.574675} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.635172] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 54dacd73-c760-49bd-9baf-ad2012a53264/54dacd73-c760-49bd-9baf-ad2012a53264.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 999.635408] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 999.635656] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-024d82f9-6fb3-4c61-aa55-16155931f792 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.642703] env[61868]: DEBUG oslo_vmware.api [None req-7b1de24a-aefe-4a43-a01d-a2b914557110 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315858, 'name': PowerOffVM_Task, 'duration_secs': 0.196328} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.643928] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b1de24a-aefe-4a43-a01d-a2b914557110 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 999.644115] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-7b1de24a-aefe-4a43-a01d-a2b914557110 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 999.644455] env[61868]: DEBUG oslo_vmware.api [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 999.644455] env[61868]: value = "task-1315860" [ 999.644455] env[61868]: _type = "Task" [ 999.644455] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.644667] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5876f480-4dca-43f5-b2b3-85e8919fdf7c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.656238] env[61868]: DEBUG oslo_vmware.api [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315860, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.719545] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-7b1de24a-aefe-4a43-a01d-a2b914557110 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 999.719796] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-7b1de24a-aefe-4a43-a01d-a2b914557110 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Deleting contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 999.720064] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b1de24a-aefe-4a43-a01d-a2b914557110 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Deleting the datastore file [datastore2] 56703761-3421-44fd-aa63-9db36c86e69c {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 999.720526] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fb222828-d3ce-4576-bdb3-eeb2930605f9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.727098] env[61868]: DEBUG oslo_vmware.api [None req-7b1de24a-aefe-4a43-a01d-a2b914557110 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 999.727098] env[61868]: value = "task-1315862" [ 999.727098] env[61868]: _type = "Task" [ 999.727098] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.735462] env[61868]: DEBUG oslo_vmware.api [None req-7b1de24a-aefe-4a43-a01d-a2b914557110 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315862, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.776660] env[61868]: DEBUG oslo_vmware.api [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52606736-412e-1bef-434a-c3dc7740cbd7, 'name': SearchDatastore_Task, 'duration_secs': 0.021444} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.777011] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.777356] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb/69a0b94b-2bf4-49ad-8fb0-ea945176e5eb.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 999.777705] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4a58b8a1-bc07-4a5b-8804-c6dab6fb8a82 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.784030] env[61868]: DEBUG oslo_vmware.api [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for the task: (returnval){ [ 999.784030] env[61868]: value = "task-1315863" [ 999.784030] env[61868]: _type = "Task" [ 999.784030] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.791986] env[61868]: DEBUG oslo_vmware.api [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315863, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.958974] env[61868]: DEBUG oslo_vmware.api [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315859, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.157410] env[61868]: DEBUG oslo_vmware.api [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315860, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065068} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.157681] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1000.158470] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bf19a32-ff65-4cca-a358-ea308f050f01 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.180613] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] 54dacd73-c760-49bd-9baf-ad2012a53264/54dacd73-c760-49bd-9baf-ad2012a53264.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1000.180900] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d1de27f5-2095-4bf9-b30d-c2f15034468e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.201459] env[61868]: DEBUG oslo_vmware.api [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 1000.201459] env[61868]: value = "task-1315864" [ 1000.201459] env[61868]: _type = "Task" [ 1000.201459] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.210023] env[61868]: DEBUG oslo_vmware.api [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315864, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.238451] env[61868]: DEBUG oslo_vmware.api [None req-7b1de24a-aefe-4a43-a01d-a2b914557110 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315862, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.391456} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.238737] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b1de24a-aefe-4a43-a01d-a2b914557110 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1000.238988] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-7b1de24a-aefe-4a43-a01d-a2b914557110 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Deleted contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1000.239306] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-7b1de24a-aefe-4a43-a01d-a2b914557110 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1000.239536] env[61868]: INFO nova.compute.manager [None req-7b1de24a-aefe-4a43-a01d-a2b914557110 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1000.239877] env[61868]: DEBUG oslo.service.loopingcall [None req-7b1de24a-aefe-4a43-a01d-a2b914557110 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1000.240029] env[61868]: DEBUG nova.compute.manager [-] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1000.240123] env[61868]: DEBUG nova.network.neutron [-] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1000.294972] env[61868]: DEBUG oslo_vmware.api [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315863, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.459563] env[61868]: DEBUG oslo_vmware.api [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315859, 'name': PowerOnVM_Task, 'duration_secs': 0.681436} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.459563] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1000.502402] env[61868]: DEBUG oslo_concurrency.lockutils [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "5d2b56d5-a4e0-4dba-9158-71036d60a40b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.502729] env[61868]: DEBUG oslo_concurrency.lockutils [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "5d2b56d5-a4e0-4dba-9158-71036d60a40b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.502919] env[61868]: DEBUG oslo_concurrency.lockutils [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "5d2b56d5-a4e0-4dba-9158-71036d60a40b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.535032] env[61868]: DEBUG nova.compute.manager [req-63f37e61-f010-4511-80bc-ae64b05ad57b req-7e84ae14-0be8-4c2f-8020-d40ed898d437 service nova] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Received event network-vif-deleted-80d0494a-2f37-48c5-b7c6-9491f5da2bc4 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1000.535362] env[61868]: INFO nova.compute.manager [req-63f37e61-f010-4511-80bc-ae64b05ad57b req-7e84ae14-0be8-4c2f-8020-d40ed898d437 service nova] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Neutron deleted interface 80d0494a-2f37-48c5-b7c6-9491f5da2bc4; detaching it from the instance and deleting it from the info cache [ 1000.535501] env[61868]: DEBUG nova.network.neutron [req-63f37e61-f010-4511-80bc-ae64b05ad57b req-7e84ae14-0be8-4c2f-8020-d40ed898d437 service nova] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.567499] env[61868]: DEBUG nova.compute.manager [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1000.568469] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2595fbcc-7ee1-4349-86a1-dc2658f5500f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.713321] env[61868]: DEBUG oslo_vmware.api [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315864, 'name': ReconfigVM_Task, 'duration_secs': 0.452351} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.713720] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Reconfigured VM instance instance-0000005e to attach disk [datastore2] 54dacd73-c760-49bd-9baf-ad2012a53264/54dacd73-c760-49bd-9baf-ad2012a53264.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1000.714612] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fc4954b5-2597-4e99-bbf3-78ae547e1a3b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.722398] env[61868]: DEBUG oslo_vmware.api [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 1000.722398] env[61868]: value = "task-1315865" [ 1000.722398] env[61868]: _type = "Task" [ 1000.722398] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.733931] env[61868]: DEBUG oslo_vmware.api [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315865, 'name': Rename_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.795510] env[61868]: DEBUG oslo_vmware.api [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315863, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.730091} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.795772] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb/69a0b94b-2bf4-49ad-8fb0-ea945176e5eb.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1000.795988] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1000.796259] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4f7677db-0951-4546-9f5f-dd66105d7f50 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.803017] env[61868]: DEBUG oslo_vmware.api [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for the task: (returnval){ [ 1000.803017] env[61868]: value = "task-1315866" [ 1000.803017] env[61868]: _type = "Task" [ 1000.803017] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.810404] env[61868]: DEBUG oslo_vmware.api [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315866, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.009255] env[61868]: DEBUG nova.network.neutron [-] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.038653] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e32f2b2b-89ed-456b-a527-f9bef319467f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.050590] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-536fb7bf-7d76-4627-9b6f-82584f9e0f4e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.084502] env[61868]: DEBUG nova.compute.manager [req-63f37e61-f010-4511-80bc-ae64b05ad57b req-7e84ae14-0be8-4c2f-8020-d40ed898d437 service nova] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Detach interface failed, port_id=80d0494a-2f37-48c5-b7c6-9491f5da2bc4, reason: Instance 56703761-3421-44fd-aa63-9db36c86e69c could not be found. {{(pid=61868) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1001.088931] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9e031c51-26c7-4584-99d8-31720c367624 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lock "75fdbc11-1ef3-442d-bcf4-55069ba9ab62" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 25.346s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.232921] env[61868]: DEBUG oslo_vmware.api [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315865, 'name': Rename_Task, 'duration_secs': 0.157921} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.233177] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1001.233439] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a03e33c0-e63f-4537-8449-49fa76d82e70 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.240975] env[61868]: DEBUG oslo_vmware.api [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 1001.240975] env[61868]: value = "task-1315867" [ 1001.240975] env[61868]: _type = "Task" [ 1001.240975] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.249379] env[61868]: DEBUG oslo_vmware.api [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315867, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.312239] env[61868]: DEBUG oslo_vmware.api [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315866, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.108449} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.312531] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1001.313324] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-737b43e1-fef7-4420-a48a-7d7eeaa49b0c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.335134] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb/69a0b94b-2bf4-49ad-8fb0-ea945176e5eb.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1001.335471] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-396393c6-596e-4a9d-8c53-3ce4d5e12cab {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.356705] env[61868]: DEBUG oslo_vmware.api [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for the task: (returnval){ [ 1001.356705] env[61868]: value = "task-1315868" [ 1001.356705] env[61868]: _type = "Task" [ 1001.356705] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.365702] env[61868]: DEBUG oslo_vmware.api [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315868, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.512227] env[61868]: INFO nova.compute.manager [-] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Took 1.27 seconds to deallocate network for instance. [ 1001.543514] env[61868]: DEBUG oslo_concurrency.lockutils [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "refresh_cache-5d2b56d5-a4e0-4dba-9158-71036d60a40b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.543737] env[61868]: DEBUG oslo_concurrency.lockutils [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquired lock "refresh_cache-5d2b56d5-a4e0-4dba-9158-71036d60a40b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.543920] env[61868]: DEBUG nova.network.neutron [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1001.752088] env[61868]: DEBUG oslo_vmware.api [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315867, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.866581] env[61868]: DEBUG oslo_vmware.api [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315868, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.018704] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7b1de24a-aefe-4a43-a01d-a2b914557110 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.019081] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7b1de24a-aefe-4a43-a01d-a2b914557110 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.019262] env[61868]: DEBUG nova.objects.instance [None req-7b1de24a-aefe-4a43-a01d-a2b914557110 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lazy-loading 'resources' on Instance uuid 56703761-3421-44fd-aa63-9db36c86e69c {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1002.253661] env[61868]: DEBUG oslo_vmware.api [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315867, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.370829] env[61868]: DEBUG oslo_vmware.api [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315868, 'name': ReconfigVM_Task, 'duration_secs': 0.511628} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.370829] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Reconfigured VM instance instance-0000005f to attach disk [datastore2] 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb/69a0b94b-2bf4-49ad-8fb0-ea945176e5eb.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1002.370829] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2c2ae418-1d4a-4b42-b466-22799c604526 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.379778] env[61868]: DEBUG oslo_vmware.api [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for the task: (returnval){ [ 1002.379778] env[61868]: value = "task-1315869" [ 1002.379778] env[61868]: _type = "Task" [ 1002.379778] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.389479] env[61868]: DEBUG oslo_vmware.api [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315869, 'name': Rename_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.460839] env[61868]: DEBUG nova.network.neutron [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Updating instance_info_cache with network_info: [{"id": "9023a9b8-ad2a-42b6-b2cd-b2cdb86dd9cb", "address": "fa:16:3e:be:bc:fe", "network": {"id": "c1f5bf03-438f-4acc-93fd-159bed7d4069", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2061742146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a60c3c20950e4156b64c3b4c61b9f0f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24b91e3b-50f1-4a16-b929-942e6b31b2d8", "external-id": "nsx-vlan-transportzone-34", "segmentation_id": 34, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9023a9b8-ad", "ovs_interfaceid": "9023a9b8-ad2a-42b6-b2cd-b2cdb86dd9cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.558552] env[61868]: DEBUG oslo_vmware.rw_handles [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52dc8fad-8f05-f77b-148c-f74b46174d8f/disk-0.vmdk. {{(pid=61868) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1002.559640] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39629f09-ecb6-4f28-8a2e-fc97bb35f065 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.565789] env[61868]: DEBUG oslo_vmware.rw_handles [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52dc8fad-8f05-f77b-148c-f74b46174d8f/disk-0.vmdk is in state: ready. {{(pid=61868) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1002.565985] env[61868]: ERROR oslo_vmware.rw_handles [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52dc8fad-8f05-f77b-148c-f74b46174d8f/disk-0.vmdk due to incomplete transfer. [ 1002.568210] env[61868]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-d4e39a6f-daca-40a8-9be1-2fe36454983b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.575330] env[61868]: DEBUG oslo_vmware.rw_handles [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52dc8fad-8f05-f77b-148c-f74b46174d8f/disk-0.vmdk. {{(pid=61868) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1002.575555] env[61868]: DEBUG nova.virt.vmwareapi.images [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Uploaded image 6732cc60-80ef-44f3-89ae-468b81ae48f9 to the Glance image server {{(pid=61868) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1002.577269] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Destroying the VM {{(pid=61868) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1002.577548] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-af64de96-6c7a-471a-ba3d-f7d4ec1bea5c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.586090] env[61868]: DEBUG oslo_vmware.api [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 1002.586090] env[61868]: value = "task-1315870" [ 1002.586090] env[61868]: _type = "Task" [ 1002.586090] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.594550] env[61868]: DEBUG oslo_vmware.api [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315870, 'name': Destroy_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.679272] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35c8e89e-1827-45d2-aad4-32b386b7fea0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.687019] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f81c443-19aa-4bc9-9e98-6c3cd640282b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.718539] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e9dd8a6-7c6e-4ab4-9f9e-bb711b6616c0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.726551] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c08bc61e-6c94-45a8-822c-2d629808c31f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.740532] env[61868]: DEBUG nova.compute.provider_tree [None req-7b1de24a-aefe-4a43-a01d-a2b914557110 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1002.750218] env[61868]: DEBUG oslo_vmware.api [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315867, 'name': PowerOnVM_Task, 'duration_secs': 1.238916} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.751965] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1002.752224] env[61868]: INFO nova.compute.manager [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Took 9.68 seconds to spawn the instance on the hypervisor. [ 1002.752474] env[61868]: DEBUG nova.compute.manager [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1002.757464] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ca5aaa1-2a4b-4bba-a5d3-e7c9dbc1b33e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.890299] env[61868]: DEBUG oslo_vmware.api [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315869, 'name': Rename_Task, 'duration_secs': 0.259099} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.890601] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1002.890843] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3121f2bb-d242-4af1-8756-6d2bc4516f3b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.896068] env[61868]: DEBUG oslo_vmware.api [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for the task: (returnval){ [ 1002.896068] env[61868]: value = "task-1315871" [ 1002.896068] env[61868]: _type = "Task" [ 1002.896068] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.903047] env[61868]: DEBUG oslo_vmware.api [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315871, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.963531] env[61868]: DEBUG oslo_concurrency.lockutils [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Releasing lock "refresh_cache-5d2b56d5-a4e0-4dba-9158-71036d60a40b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.096026] env[61868]: DEBUG oslo_vmware.api [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315870, 'name': Destroy_Task, 'duration_secs': 0.299395} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.096329] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Destroyed the VM [ 1003.096576] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Deleting Snapshot of the VM instance {{(pid=61868) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1003.096820] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-a7335776-7e20-4e0b-8e61-0425d594c748 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.103230] env[61868]: DEBUG oslo_vmware.api [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 1003.103230] env[61868]: value = "task-1315872" [ 1003.103230] env[61868]: _type = "Task" [ 1003.103230] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.111250] env[61868]: DEBUG oslo_vmware.api [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315872, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.244435] env[61868]: DEBUG nova.scheduler.client.report [None req-7b1de24a-aefe-4a43-a01d-a2b914557110 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1003.263848] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f7dea81-c513-40cf-aed0-2b3a98e42f30 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.275295] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-e83b6507-b0e5-4e17-a128-5c41089b1933 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Suspending the VM {{(pid=61868) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1003.276480] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-95a51625-b647-496b-ae42-76885dc99b99 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.278949] env[61868]: INFO nova.compute.manager [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Took 21.70 seconds to build instance. [ 1003.284941] env[61868]: DEBUG oslo_vmware.api [None req-e83b6507-b0e5-4e17-a128-5c41089b1933 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the task: (returnval){ [ 1003.284941] env[61868]: value = "task-1315873" [ 1003.284941] env[61868]: _type = "Task" [ 1003.284941] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.293050] env[61868]: DEBUG oslo_vmware.api [None req-e83b6507-b0e5-4e17-a128-5c41089b1933 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315873, 'name': SuspendVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.406821] env[61868]: DEBUG oslo_vmware.api [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315871, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.487363] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aae7b8f-4173-4744-93e0-53c2bd74f66c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.506809] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad0263b8-2696-4e4c-8b0a-83713a770fb3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.514628] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Updating instance '5d2b56d5-a4e0-4dba-9158-71036d60a40b' progress to 83 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1003.613657] env[61868]: DEBUG oslo_vmware.api [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315872, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.749407] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7b1de24a-aefe-4a43-a01d-a2b914557110 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.730s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.769948] env[61868]: INFO nova.scheduler.client.report [None req-7b1de24a-aefe-4a43-a01d-a2b914557110 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Deleted allocations for instance 56703761-3421-44fd-aa63-9db36c86e69c [ 1003.782236] env[61868]: DEBUG oslo_concurrency.lockutils [None req-35fec6d8-7a02-488a-b6a3-54e85dbcd8aa tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "54dacd73-c760-49bd-9baf-ad2012a53264" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.215s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.795623] env[61868]: DEBUG oslo_vmware.api [None req-e83b6507-b0e5-4e17-a128-5c41089b1933 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315873, 'name': SuspendVM_Task} progress is 58%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.906833] env[61868]: DEBUG oslo_vmware.api [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315871, 'name': PowerOnVM_Task, 'duration_secs': 0.899309} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.907167] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1003.907336] env[61868]: INFO nova.compute.manager [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Took 8.53 seconds to spawn the instance on the hypervisor. [ 1003.907510] env[61868]: DEBUG nova.compute.manager [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1003.908300] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3ee3a1a-6315-42d1-ac80-ceeb5ebcec05 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.021320] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1004.021619] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-372524d4-9eb3-46be-980f-366236f21fac {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.028516] env[61868]: DEBUG oslo_vmware.api [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 1004.028516] env[61868]: value = "task-1315874" [ 1004.028516] env[61868]: _type = "Task" [ 1004.028516] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.036069] env[61868]: DEBUG oslo_vmware.api [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315874, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.113752] env[61868]: DEBUG oslo_vmware.api [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315872, 'name': RemoveSnapshot_Task, 'duration_secs': 0.717101} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.113959] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Deleted Snapshot of the VM instance {{(pid=61868) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1004.114201] env[61868]: INFO nova.compute.manager [None req-2caac5c8-3ad8-42e1-9eaa-e846b4331968 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Took 18.01 seconds to snapshot the instance on the hypervisor. [ 1004.277639] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7b1de24a-aefe-4a43-a01d-a2b914557110 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "56703761-3421-44fd-aa63-9db36c86e69c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.176s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.296155] env[61868]: DEBUG oslo_vmware.api [None req-e83b6507-b0e5-4e17-a128-5c41089b1933 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315873, 'name': SuspendVM_Task} progress is 100%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.428690] env[61868]: INFO nova.compute.manager [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Took 19.40 seconds to build instance. [ 1004.509380] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "c96c4bcb-03f9-4961-aa17-8364aa4cf66e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.509380] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "c96c4bcb-03f9-4961-aa17-8364aa4cf66e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.538631] env[61868]: DEBUG oslo_vmware.api [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315874, 'name': PowerOnVM_Task, 'duration_secs': 0.38398} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.539190] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1004.539520] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-53ddd3a1-26cf-4b63-a1ee-2d6fdcf9eb58 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Updating instance '5d2b56d5-a4e0-4dba-9158-71036d60a40b' progress to 100 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1004.594952] env[61868]: DEBUG nova.compute.manager [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Stashing vm_state: active {{(pid=61868) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1004.796608] env[61868]: DEBUG oslo_vmware.api [None req-e83b6507-b0e5-4e17-a128-5c41089b1933 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315873, 'name': SuspendVM_Task, 'duration_secs': 1.04357} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.796968] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-e83b6507-b0e5-4e17-a128-5c41089b1933 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Suspended the VM {{(pid=61868) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1004.797126] env[61868]: DEBUG nova.compute.manager [None req-e83b6507-b0e5-4e17-a128-5c41089b1933 tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1004.797985] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5990c216-02c7-4696-85c8-915e743fc7ff {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.931403] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5c89c380-7633-4ce9-943d-3e48b259f6cb tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "69a0b94b-2bf4-49ad-8fb0-ea945176e5eb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.908s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.011671] env[61868]: DEBUG nova.compute.manager [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1005.118346] env[61868]: DEBUG oslo_concurrency.lockutils [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.118578] env[61868]: DEBUG oslo_concurrency.lockutils [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.327801] env[61868]: DEBUG oslo_concurrency.lockutils [None req-86c46a89-abde-41f6-8053-5e092f012a9a tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "9286cc76-1152-4fd4-9279-33d6ee4b9486" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.327801] env[61868]: DEBUG oslo_concurrency.lockutils [None req-86c46a89-abde-41f6-8053-5e092f012a9a tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "9286cc76-1152-4fd4-9279-33d6ee4b9486" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.327801] env[61868]: DEBUG oslo_concurrency.lockutils [None req-86c46a89-abde-41f6-8053-5e092f012a9a tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "9286cc76-1152-4fd4-9279-33d6ee4b9486-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.327801] env[61868]: DEBUG oslo_concurrency.lockutils [None req-86c46a89-abde-41f6-8053-5e092f012a9a tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "9286cc76-1152-4fd4-9279-33d6ee4b9486-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.327801] env[61868]: DEBUG oslo_concurrency.lockutils [None req-86c46a89-abde-41f6-8053-5e092f012a9a tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "9286cc76-1152-4fd4-9279-33d6ee4b9486-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.329034] env[61868]: INFO nova.compute.manager [None req-86c46a89-abde-41f6-8053-5e092f012a9a tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Terminating instance [ 1005.330944] env[61868]: DEBUG nova.compute.manager [None req-86c46a89-abde-41f6-8053-5e092f012a9a tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1005.331169] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-86c46a89-abde-41f6-8053-5e092f012a9a tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1005.332022] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c4a247f-f3b9-4735-a5e0-57db108cc545 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.340236] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-86c46a89-abde-41f6-8053-5e092f012a9a tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1005.340499] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f922c73f-5513-4f7e-aff8-4d1b03ce7878 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.347602] env[61868]: DEBUG oslo_vmware.api [None req-86c46a89-abde-41f6-8053-5e092f012a9a tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 1005.347602] env[61868]: value = "task-1315875" [ 1005.347602] env[61868]: _type = "Task" [ 1005.347602] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.355980] env[61868]: DEBUG oslo_vmware.api [None req-86c46a89-abde-41f6-8053-5e092f012a9a tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315875, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.407966] env[61868]: DEBUG nova.compute.manager [req-69fa1458-6afa-4424-bb9b-861f16ba60f3 req-21800f9f-3ecf-4a9c-82fc-ba6557d3337d service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Received event network-changed-cd9efdc6-e6f5-4449-b95d-d1fbec53a295 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1005.408297] env[61868]: DEBUG nova.compute.manager [req-69fa1458-6afa-4424-bb9b-861f16ba60f3 req-21800f9f-3ecf-4a9c-82fc-ba6557d3337d service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Refreshing instance network info cache due to event network-changed-cd9efdc6-e6f5-4449-b95d-d1fbec53a295. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1005.408610] env[61868]: DEBUG oslo_concurrency.lockutils [req-69fa1458-6afa-4424-bb9b-861f16ba60f3 req-21800f9f-3ecf-4a9c-82fc-ba6557d3337d service nova] Acquiring lock "refresh_cache-a96cab5a-c053-4e45-96f6-2aba0a819110" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.408860] env[61868]: DEBUG oslo_concurrency.lockutils [req-69fa1458-6afa-4424-bb9b-861f16ba60f3 req-21800f9f-3ecf-4a9c-82fc-ba6557d3337d service nova] Acquired lock "refresh_cache-a96cab5a-c053-4e45-96f6-2aba0a819110" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.409143] env[61868]: DEBUG nova.network.neutron [req-69fa1458-6afa-4424-bb9b-861f16ba60f3 req-21800f9f-3ecf-4a9c-82fc-ba6557d3337d service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Refreshing network info cache for port cd9efdc6-e6f5-4449-b95d-d1fbec53a295 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1005.537459] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.624428] env[61868]: INFO nova.compute.claims [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1005.655014] env[61868]: INFO nova.compute.manager [None req-9cc7a6f1-a436-4fa8-b5c3-8f56d25a815c tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Resuming [ 1005.655679] env[61868]: DEBUG nova.objects.instance [None req-9cc7a6f1-a436-4fa8-b5c3-8f56d25a815c tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lazy-loading 'flavor' on Instance uuid 75fdbc11-1ef3-442d-bcf4-55069ba9ab62 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1005.858006] env[61868]: DEBUG oslo_vmware.api [None req-86c46a89-abde-41f6-8053-5e092f012a9a tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315875, 'name': PowerOffVM_Task, 'duration_secs': 0.346789} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.858363] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-86c46a89-abde-41f6-8053-5e092f012a9a tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1005.858541] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-86c46a89-abde-41f6-8053-5e092f012a9a tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1005.858826] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0f89bcf4-1045-49c7-b7c7-f6a4a97fc31d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.131183] env[61868]: INFO nova.compute.resource_tracker [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Updating resource usage from migration 8b2b7aa9-6eb5-4898-bdc4-cd3c2b1acdd8 [ 1006.231029] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-86c46a89-abde-41f6-8053-5e092f012a9a tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1006.231029] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-86c46a89-abde-41f6-8053-5e092f012a9a tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1006.231029] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-86c46a89-abde-41f6-8053-5e092f012a9a tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Deleting the datastore file [datastore1] 9286cc76-1152-4fd4-9279-33d6ee4b9486 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1006.231640] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ecf129b8-4c8f-49a1-885f-e3b5e22f49ce {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.239693] env[61868]: DEBUG oslo_vmware.api [None req-86c46a89-abde-41f6-8053-5e092f012a9a tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 1006.239693] env[61868]: value = "task-1315877" [ 1006.239693] env[61868]: _type = "Task" [ 1006.239693] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.255913] env[61868]: DEBUG oslo_vmware.api [None req-86c46a89-abde-41f6-8053-5e092f012a9a tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315877, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.270387] env[61868]: DEBUG nova.network.neutron [req-69fa1458-6afa-4424-bb9b-861f16ba60f3 req-21800f9f-3ecf-4a9c-82fc-ba6557d3337d service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Updated VIF entry in instance network info cache for port cd9efdc6-e6f5-4449-b95d-d1fbec53a295. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1006.270387] env[61868]: DEBUG nova.network.neutron [req-69fa1458-6afa-4424-bb9b-861f16ba60f3 req-21800f9f-3ecf-4a9c-82fc-ba6557d3337d service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Updating instance_info_cache with network_info: [{"id": "cd9efdc6-e6f5-4449-b95d-d1fbec53a295", "address": "fa:16:3e:bf:13:19", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd9efdc6-e6", "ovs_interfaceid": "cd9efdc6-e6f5-4449-b95d-d1fbec53a295", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.366652] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed7111b4-c273-4af7-84f4-881a53e1e507 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.375890] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eafa5986-5084-4155-b619-8afae92e5c0a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.408601] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac7fba06-dd05-4736-8448-5fae98656766 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.416939] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7e96dbb-e3f3-416f-929f-e164a479ae39 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.430726] env[61868]: DEBUG nova.compute.provider_tree [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1006.533645] env[61868]: DEBUG oslo_concurrency.lockutils [None req-edf96371-ba85-4522-b84c-d581f28faebc tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "5d2b56d5-a4e0-4dba-9158-71036d60a40b" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.534008] env[61868]: DEBUG oslo_concurrency.lockutils [None req-edf96371-ba85-4522-b84c-d581f28faebc tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "5d2b56d5-a4e0-4dba-9158-71036d60a40b" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.534285] env[61868]: DEBUG nova.compute.manager [None req-edf96371-ba85-4522-b84c-d581f28faebc tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Going to confirm migration 2 {{(pid=61868) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1006.665757] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9cc7a6f1-a436-4fa8-b5c3-8f56d25a815c tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquiring lock "refresh_cache-75fdbc11-1ef3-442d-bcf4-55069ba9ab62" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.665997] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9cc7a6f1-a436-4fa8-b5c3-8f56d25a815c tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquired lock "refresh_cache-75fdbc11-1ef3-442d-bcf4-55069ba9ab62" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.666354] env[61868]: DEBUG nova.network.neutron [None req-9cc7a6f1-a436-4fa8-b5c3-8f56d25a815c tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1006.752976] env[61868]: DEBUG oslo_vmware.api [None req-86c46a89-abde-41f6-8053-5e092f012a9a tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315877, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.25273} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.753799] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-86c46a89-abde-41f6-8053-5e092f012a9a tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1006.753799] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-86c46a89-abde-41f6-8053-5e092f012a9a tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1006.753799] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-86c46a89-abde-41f6-8053-5e092f012a9a tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1006.754040] env[61868]: INFO nova.compute.manager [None req-86c46a89-abde-41f6-8053-5e092f012a9a tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Took 1.42 seconds to destroy the instance on the hypervisor. [ 1006.754238] env[61868]: DEBUG oslo.service.loopingcall [None req-86c46a89-abde-41f6-8053-5e092f012a9a tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1006.754436] env[61868]: DEBUG nova.compute.manager [-] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1006.754534] env[61868]: DEBUG nova.network.neutron [-] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1006.773111] env[61868]: DEBUG oslo_concurrency.lockutils [req-69fa1458-6afa-4424-bb9b-861f16ba60f3 req-21800f9f-3ecf-4a9c-82fc-ba6557d3337d service nova] Releasing lock "refresh_cache-a96cab5a-c053-4e45-96f6-2aba0a819110" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.934623] env[61868]: DEBUG nova.scheduler.client.report [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1007.036810] env[61868]: DEBUG nova.compute.manager [req-dba87363-dff9-477c-8b5f-15052b0849fa req-17a11270-747c-421d-8e4e-55374a7414c9 service nova] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Received event network-vif-deleted-4edbd65b-1856-47a2-95c9-e4c8f4d04c71 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1007.037045] env[61868]: INFO nova.compute.manager [req-dba87363-dff9-477c-8b5f-15052b0849fa req-17a11270-747c-421d-8e4e-55374a7414c9 service nova] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Neutron deleted interface 4edbd65b-1856-47a2-95c9-e4c8f4d04c71; detaching it from the instance and deleting it from the info cache [ 1007.037229] env[61868]: DEBUG nova.network.neutron [req-dba87363-dff9-477c-8b5f-15052b0849fa req-17a11270-747c-421d-8e4e-55374a7414c9 service nova] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.102946] env[61868]: DEBUG oslo_concurrency.lockutils [None req-edf96371-ba85-4522-b84c-d581f28faebc tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "refresh_cache-5d2b56d5-a4e0-4dba-9158-71036d60a40b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1007.103191] env[61868]: DEBUG oslo_concurrency.lockutils [None req-edf96371-ba85-4522-b84c-d581f28faebc tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquired lock "refresh_cache-5d2b56d5-a4e0-4dba-9158-71036d60a40b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.103376] env[61868]: DEBUG nova.network.neutron [None req-edf96371-ba85-4522-b84c-d581f28faebc tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1007.103589] env[61868]: DEBUG nova.objects.instance [None req-edf96371-ba85-4522-b84c-d581f28faebc tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lazy-loading 'info_cache' on Instance uuid 5d2b56d5-a4e0-4dba-9158-71036d60a40b {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1007.440149] env[61868]: DEBUG nova.compute.manager [req-d985dcc5-c160-4b46-a5fd-b66f8249c2bc req-fc590693-8a6d-4df6-ba8a-9559089627c3 service nova] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Received event network-changed-f8f4c521-03cd-4e54-99f1-761d9a515aa3 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1007.440149] env[61868]: DEBUG nova.compute.manager [req-d985dcc5-c160-4b46-a5fd-b66f8249c2bc req-fc590693-8a6d-4df6-ba8a-9559089627c3 service nova] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Refreshing instance network info cache due to event network-changed-f8f4c521-03cd-4e54-99f1-761d9a515aa3. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1007.440149] env[61868]: DEBUG oslo_concurrency.lockutils [req-d985dcc5-c160-4b46-a5fd-b66f8249c2bc req-fc590693-8a6d-4df6-ba8a-9559089627c3 service nova] Acquiring lock "refresh_cache-69a0b94b-2bf4-49ad-8fb0-ea945176e5eb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1007.440149] env[61868]: DEBUG oslo_concurrency.lockutils [req-d985dcc5-c160-4b46-a5fd-b66f8249c2bc req-fc590693-8a6d-4df6-ba8a-9559089627c3 service nova] Acquired lock "refresh_cache-69a0b94b-2bf4-49ad-8fb0-ea945176e5eb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.440149] env[61868]: DEBUG nova.network.neutron [req-d985dcc5-c160-4b46-a5fd-b66f8249c2bc req-fc590693-8a6d-4df6-ba8a-9559089627c3 service nova] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Refreshing network info cache for port f8f4c521-03cd-4e54-99f1-761d9a515aa3 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1007.443658] env[61868]: DEBUG oslo_concurrency.lockutils [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.325s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.443901] env[61868]: INFO nova.compute.manager [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Migrating [ 1007.452417] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.915s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.453980] env[61868]: INFO nova.compute.claims [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1007.481585] env[61868]: DEBUG nova.network.neutron [None req-9cc7a6f1-a436-4fa8-b5c3-8f56d25a815c tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Updating instance_info_cache with network_info: [{"id": "21d6cab8-f691-425c-bfa8-9d2f1c294b64", "address": "fa:16:3e:76:a4:70", "network": {"id": "e7457f1a-171a-4c82-a260-1d1a7a95861e", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1032778554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b9b800086764b41a8817d1d8e463a2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21d6cab8-f6", "ovs_interfaceid": "21d6cab8-f691-425c-bfa8-9d2f1c294b64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.513666] env[61868]: DEBUG nova.network.neutron [-] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.543865] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f2f9752b-87e9-483a-ab1d-ba14d7db3979 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.553688] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9efc95c-7506-4c59-915a-237707dbfc2b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.582198] env[61868]: DEBUG nova.compute.manager [req-dba87363-dff9-477c-8b5f-15052b0849fa req-17a11270-747c-421d-8e4e-55374a7414c9 service nova] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Detach interface failed, port_id=4edbd65b-1856-47a2-95c9-e4c8f4d04c71, reason: Instance 9286cc76-1152-4fd4-9279-33d6ee4b9486 could not be found. {{(pid=61868) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1007.680362] env[61868]: DEBUG nova.network.neutron [req-d985dcc5-c160-4b46-a5fd-b66f8249c2bc req-fc590693-8a6d-4df6-ba8a-9559089627c3 service nova] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Updated VIF entry in instance network info cache for port f8f4c521-03cd-4e54-99f1-761d9a515aa3. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1007.680716] env[61868]: DEBUG nova.network.neutron [req-d985dcc5-c160-4b46-a5fd-b66f8249c2bc req-fc590693-8a6d-4df6-ba8a-9559089627c3 service nova] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Updating instance_info_cache with network_info: [{"id": "f8f4c521-03cd-4e54-99f1-761d9a515aa3", "address": "fa:16:3e:3c:dd:65", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8f4c521-03", "ovs_interfaceid": "f8f4c521-03cd-4e54-99f1-761d9a515aa3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.965307] env[61868]: DEBUG oslo_concurrency.lockutils [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "refresh_cache-54dacd73-c760-49bd-9baf-ad2012a53264" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1007.965573] env[61868]: DEBUG oslo_concurrency.lockutils [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquired lock "refresh_cache-54dacd73-c760-49bd-9baf-ad2012a53264" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.965656] env[61868]: DEBUG nova.network.neutron [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1007.983722] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9cc7a6f1-a436-4fa8-b5c3-8f56d25a815c tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Releasing lock "refresh_cache-75fdbc11-1ef3-442d-bcf4-55069ba9ab62" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.984660] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97bfa61e-b28c-4181-80f5-a504bfb3b3b9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.992144] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9cc7a6f1-a436-4fa8-b5c3-8f56d25a815c tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Resuming the VM {{(pid=61868) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1007.992144] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3e876422-d4d8-4f5f-9b10-ba74fb5968ae {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.997685] env[61868]: DEBUG oslo_vmware.api [None req-9cc7a6f1-a436-4fa8-b5c3-8f56d25a815c tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the task: (returnval){ [ 1007.997685] env[61868]: value = "task-1315878" [ 1007.997685] env[61868]: _type = "Task" [ 1007.997685] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.004918] env[61868]: DEBUG oslo_vmware.api [None req-9cc7a6f1-a436-4fa8-b5c3-8f56d25a815c tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315878, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.017551] env[61868]: INFO nova.compute.manager [-] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Took 1.26 seconds to deallocate network for instance. [ 1008.185625] env[61868]: DEBUG oslo_concurrency.lockutils [req-d985dcc5-c160-4b46-a5fd-b66f8249c2bc req-fc590693-8a6d-4df6-ba8a-9559089627c3 service nova] Releasing lock "refresh_cache-69a0b94b-2bf4-49ad-8fb0-ea945176e5eb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.185902] env[61868]: DEBUG nova.compute.manager [req-d985dcc5-c160-4b46-a5fd-b66f8249c2bc req-fc590693-8a6d-4df6-ba8a-9559089627c3 service nova] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Received event network-changed-f8f4c521-03cd-4e54-99f1-761d9a515aa3 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1008.186087] env[61868]: DEBUG nova.compute.manager [req-d985dcc5-c160-4b46-a5fd-b66f8249c2bc req-fc590693-8a6d-4df6-ba8a-9559089627c3 service nova] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Refreshing instance network info cache due to event network-changed-f8f4c521-03cd-4e54-99f1-761d9a515aa3. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1008.186298] env[61868]: DEBUG oslo_concurrency.lockutils [req-d985dcc5-c160-4b46-a5fd-b66f8249c2bc req-fc590693-8a6d-4df6-ba8a-9559089627c3 service nova] Acquiring lock "refresh_cache-69a0b94b-2bf4-49ad-8fb0-ea945176e5eb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.186459] env[61868]: DEBUG oslo_concurrency.lockutils [req-d985dcc5-c160-4b46-a5fd-b66f8249c2bc req-fc590693-8a6d-4df6-ba8a-9559089627c3 service nova] Acquired lock "refresh_cache-69a0b94b-2bf4-49ad-8fb0-ea945176e5eb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.186627] env[61868]: DEBUG nova.network.neutron [req-d985dcc5-c160-4b46-a5fd-b66f8249c2bc req-fc590693-8a6d-4df6-ba8a-9559089627c3 service nova] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Refreshing network info cache for port f8f4c521-03cd-4e54-99f1-761d9a515aa3 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1008.309895] env[61868]: DEBUG nova.network.neutron [None req-edf96371-ba85-4522-b84c-d581f28faebc tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Updating instance_info_cache with network_info: [{"id": "9023a9b8-ad2a-42b6-b2cd-b2cdb86dd9cb", "address": "fa:16:3e:be:bc:fe", "network": {"id": "c1f5bf03-438f-4acc-93fd-159bed7d4069", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2061742146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a60c3c20950e4156b64c3b4c61b9f0f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24b91e3b-50f1-4a16-b929-942e6b31b2d8", "external-id": "nsx-vlan-transportzone-34", "segmentation_id": 34, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9023a9b8-ad", "ovs_interfaceid": "9023a9b8-ad2a-42b6-b2cd-b2cdb86dd9cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.507690] env[61868]: DEBUG oslo_vmware.api [None req-9cc7a6f1-a436-4fa8-b5c3-8f56d25a815c tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315878, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.525904] env[61868]: DEBUG oslo_concurrency.lockutils [None req-86c46a89-abde-41f6-8053-5e092f012a9a tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.669122] env[61868]: DEBUG nova.network.neutron [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Updating instance_info_cache with network_info: [{"id": "cbd81a07-4122-41bd-bb91-d47b6bc3918d", "address": "fa:16:3e:31:4b:86", "network": {"id": "5b4c5906-b873-4369-855b-e0898ad84812", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-254512637-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "812e115cc1114d0e8536924c4f5556f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcbd81a07-41", "ovs_interfaceid": "cbd81a07-4122-41bd-bb91-d47b6bc3918d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.683909] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44d5b831-1f8c-4d5d-a15c-b808262260bf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.693193] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-993c93af-6861-449c-a1e7-0f161fd989da {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.724899] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99e343ce-7b76-4390-b014-d45d456d08a8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.731981] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aa84a74-9e20-4a50-85d6-4c69d978d4c5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.744641] env[61868]: DEBUG nova.compute.provider_tree [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1008.812254] env[61868]: DEBUG oslo_concurrency.lockutils [None req-edf96371-ba85-4522-b84c-d581f28faebc tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Releasing lock "refresh_cache-5d2b56d5-a4e0-4dba-9158-71036d60a40b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.812642] env[61868]: DEBUG nova.objects.instance [None req-edf96371-ba85-4522-b84c-d581f28faebc tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lazy-loading 'migration_context' on Instance uuid 5d2b56d5-a4e0-4dba-9158-71036d60a40b {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1008.879283] env[61868]: DEBUG nova.network.neutron [req-d985dcc5-c160-4b46-a5fd-b66f8249c2bc req-fc590693-8a6d-4df6-ba8a-9559089627c3 service nova] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Updated VIF entry in instance network info cache for port f8f4c521-03cd-4e54-99f1-761d9a515aa3. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1008.879640] env[61868]: DEBUG nova.network.neutron [req-d985dcc5-c160-4b46-a5fd-b66f8249c2bc req-fc590693-8a6d-4df6-ba8a-9559089627c3 service nova] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Updating instance_info_cache with network_info: [{"id": "f8f4c521-03cd-4e54-99f1-761d9a515aa3", "address": "fa:16:3e:3c:dd:65", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8f4c521-03", "ovs_interfaceid": "f8f4c521-03cd-4e54-99f1-761d9a515aa3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.008786] env[61868]: DEBUG oslo_vmware.api [None req-9cc7a6f1-a436-4fa8-b5c3-8f56d25a815c tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315878, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.171838] env[61868]: DEBUG oslo_concurrency.lockutils [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Releasing lock "refresh_cache-54dacd73-c760-49bd-9baf-ad2012a53264" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.248849] env[61868]: DEBUG nova.scheduler.client.report [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1009.315285] env[61868]: DEBUG nova.objects.base [None req-edf96371-ba85-4522-b84c-d581f28faebc tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Object Instance<5d2b56d5-a4e0-4dba-9158-71036d60a40b> lazy-loaded attributes: info_cache,migration_context {{(pid=61868) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1009.316282] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab53fa70-11c7-4fb0-9a6b-e6a85d7770cb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.339376] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4a037f4-10e1-43f6-8c0c-2828dc8eeae1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.348119] env[61868]: DEBUG oslo_vmware.api [None req-edf96371-ba85-4522-b84c-d581f28faebc tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 1009.348119] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5211272f-21e4-a41e-4a62-14542ae49fff" [ 1009.348119] env[61868]: _type = "Task" [ 1009.348119] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.358737] env[61868]: DEBUG oslo_vmware.api [None req-edf96371-ba85-4522-b84c-d581f28faebc tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5211272f-21e4-a41e-4a62-14542ae49fff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.381773] env[61868]: DEBUG oslo_concurrency.lockutils [req-d985dcc5-c160-4b46-a5fd-b66f8249c2bc req-fc590693-8a6d-4df6-ba8a-9559089627c3 service nova] Releasing lock "refresh_cache-69a0b94b-2bf4-49ad-8fb0-ea945176e5eb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.382093] env[61868]: DEBUG nova.compute.manager [req-d985dcc5-c160-4b46-a5fd-b66f8249c2bc req-fc590693-8a6d-4df6-ba8a-9559089627c3 service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Received event network-changed-cd9efdc6-e6f5-4449-b95d-d1fbec53a295 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1009.382278] env[61868]: DEBUG nova.compute.manager [req-d985dcc5-c160-4b46-a5fd-b66f8249c2bc req-fc590693-8a6d-4df6-ba8a-9559089627c3 service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Refreshing instance network info cache due to event network-changed-cd9efdc6-e6f5-4449-b95d-d1fbec53a295. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1009.382568] env[61868]: DEBUG oslo_concurrency.lockutils [req-d985dcc5-c160-4b46-a5fd-b66f8249c2bc req-fc590693-8a6d-4df6-ba8a-9559089627c3 service nova] Acquiring lock "refresh_cache-a96cab5a-c053-4e45-96f6-2aba0a819110" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.382716] env[61868]: DEBUG oslo_concurrency.lockutils [req-d985dcc5-c160-4b46-a5fd-b66f8249c2bc req-fc590693-8a6d-4df6-ba8a-9559089627c3 service nova] Acquired lock "refresh_cache-a96cab5a-c053-4e45-96f6-2aba0a819110" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.382882] env[61868]: DEBUG nova.network.neutron [req-d985dcc5-c160-4b46-a5fd-b66f8249c2bc req-fc590693-8a6d-4df6-ba8a-9559089627c3 service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Refreshing network info cache for port cd9efdc6-e6f5-4449-b95d-d1fbec53a295 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1009.510369] env[61868]: DEBUG oslo_vmware.api [None req-9cc7a6f1-a436-4fa8-b5c3-8f56d25a815c tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315878, 'name': PowerOnVM_Task, 'duration_secs': 1.328787} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.510638] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9cc7a6f1-a436-4fa8-b5c3-8f56d25a815c tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Resumed the VM {{(pid=61868) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1009.510821] env[61868]: DEBUG nova.compute.manager [None req-9cc7a6f1-a436-4fa8-b5c3-8f56d25a815c tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1009.511616] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-961f7478-7167-4f94-a79c-3eb094991d2d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.760371] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.308s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.760937] env[61868]: DEBUG nova.compute.manager [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1009.763462] env[61868]: DEBUG oslo_concurrency.lockutils [None req-86c46a89-abde-41f6-8053-5e092f012a9a tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.238s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.763726] env[61868]: DEBUG nova.objects.instance [None req-86c46a89-abde-41f6-8053-5e092f012a9a tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lazy-loading 'resources' on Instance uuid 9286cc76-1152-4fd4-9279-33d6ee4b9486 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1009.858985] env[61868]: DEBUG oslo_vmware.api [None req-edf96371-ba85-4522-b84c-d581f28faebc tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5211272f-21e4-a41e-4a62-14542ae49fff, 'name': SearchDatastore_Task, 'duration_secs': 0.011612} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.858985] env[61868]: DEBUG oslo_concurrency.lockutils [None req-edf96371-ba85-4522-b84c-d581f28faebc tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.114834] env[61868]: DEBUG nova.network.neutron [req-d985dcc5-c160-4b46-a5fd-b66f8249c2bc req-fc590693-8a6d-4df6-ba8a-9559089627c3 service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Updated VIF entry in instance network info cache for port cd9efdc6-e6f5-4449-b95d-d1fbec53a295. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1010.115406] env[61868]: DEBUG nova.network.neutron [req-d985dcc5-c160-4b46-a5fd-b66f8249c2bc req-fc590693-8a6d-4df6-ba8a-9559089627c3 service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Updating instance_info_cache with network_info: [{"id": "cd9efdc6-e6f5-4449-b95d-d1fbec53a295", "address": "fa:16:3e:bf:13:19", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd9efdc6-e6", "ovs_interfaceid": "cd9efdc6-e6f5-4449-b95d-d1fbec53a295", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.267160] env[61868]: DEBUG nova.compute.utils [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1010.269027] env[61868]: DEBUG nova.compute.manager [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1010.269027] env[61868]: DEBUG nova.network.neutron [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1010.313809] env[61868]: DEBUG nova.policy [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0b615dd3d1064706bea31b52fbfa9a61', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5658848e1b0e42929c04a4a8de40a291', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 1010.449140] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e989de2-23d3-47c0-8296-abdb80cdb7d5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.457586] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f03f6823-83b1-48ca-9460-87f183bc8565 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.487938] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92b5acae-c342-420f-ae42-38ae59be2f95 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.495399] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cb4f63b-8061-46f0-bfb8-f217422ffc2e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.509419] env[61868]: DEBUG nova.compute.provider_tree [None req-86c46a89-abde-41f6-8053-5e092f012a9a tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1010.598242] env[61868]: DEBUG nova.network.neutron [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Successfully created port: b0fdd6cb-31bd-4a5d-95cd-64d82af61df9 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1010.618229] env[61868]: DEBUG oslo_concurrency.lockutils [req-d985dcc5-c160-4b46-a5fd-b66f8249c2bc req-fc590693-8a6d-4df6-ba8a-9559089627c3 service nova] Releasing lock "refresh_cache-a96cab5a-c053-4e45-96f6-2aba0a819110" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.688602] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ef7b706-7860-421a-9cda-fc7bf0b449f9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.714720] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Updating instance '54dacd73-c760-49bd-9baf-ad2012a53264' progress to 0 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1010.774291] env[61868]: DEBUG nova.compute.manager [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1011.013633] env[61868]: DEBUG nova.scheduler.client.report [None req-86c46a89-abde-41f6-8053-5e092f012a9a tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1011.223257] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1011.223630] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fbfc05c3-2116-4fb4-8c9e-734398229feb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.232034] env[61868]: DEBUG oslo_vmware.api [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 1011.232034] env[61868]: value = "task-1315879" [ 1011.232034] env[61868]: _type = "Task" [ 1011.232034] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.239922] env[61868]: DEBUG oslo_vmware.api [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315879, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.519146] env[61868]: DEBUG oslo_concurrency.lockutils [None req-86c46a89-abde-41f6-8053-5e092f012a9a tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.756s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.522087] env[61868]: DEBUG oslo_concurrency.lockutils [None req-edf96371-ba85-4522-b84c-d581f28faebc tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 1.663s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.543764] env[61868]: INFO nova.scheduler.client.report [None req-86c46a89-abde-41f6-8053-5e092f012a9a tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Deleted allocations for instance 9286cc76-1152-4fd4-9279-33d6ee4b9486 [ 1011.742665] env[61868]: DEBUG oslo_vmware.api [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315879, 'name': PowerOffVM_Task, 'duration_secs': 0.175707} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.742984] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1011.743278] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Updating instance '54dacd73-c760-49bd-9baf-ad2012a53264' progress to 17 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1011.785032] env[61868]: DEBUG nova.compute.manager [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1011.824756] env[61868]: DEBUG nova.virt.hardware [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='5430e0536cce7c0b36326e4dce230a71',container_format='bare',created_at=2024-10-18T17:26:33Z,direct_url=,disk_format='vmdk',id=6732cc60-80ef-44f3-89ae-468b81ae48f9,min_disk=1,min_ram=0,name='tempest-test-snap-1744616895',owner='5658848e1b0e42929c04a4a8de40a291',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-18T17:26:51Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1011.824943] env[61868]: DEBUG nova.virt.hardware [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1011.825053] env[61868]: DEBUG nova.virt.hardware [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1011.825461] env[61868]: DEBUG nova.virt.hardware [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1011.825461] env[61868]: DEBUG nova.virt.hardware [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1011.825664] env[61868]: DEBUG nova.virt.hardware [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1011.825735] env[61868]: DEBUG nova.virt.hardware [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1011.825893] env[61868]: DEBUG nova.virt.hardware [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1011.826078] env[61868]: DEBUG nova.virt.hardware [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1011.826248] env[61868]: DEBUG nova.virt.hardware [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1011.826422] env[61868]: DEBUG nova.virt.hardware [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1011.827759] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-537a2b5b-e0a8-49ef-86d8-a560c2861f2f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.836897] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e66c0fe-6c4b-40d8-a85e-716a11bd2f05 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.991715] env[61868]: DEBUG nova.compute.manager [req-76672f56-cbaf-44b6-8b80-61a1cedeeb2f req-cc36b601-5efb-4dcd-ac2c-677119a99b85 service nova] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Received event network-vif-plugged-b0fdd6cb-31bd-4a5d-95cd-64d82af61df9 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1011.991989] env[61868]: DEBUG oslo_concurrency.lockutils [req-76672f56-cbaf-44b6-8b80-61a1cedeeb2f req-cc36b601-5efb-4dcd-ac2c-677119a99b85 service nova] Acquiring lock "c96c4bcb-03f9-4961-aa17-8364aa4cf66e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.992259] env[61868]: DEBUG oslo_concurrency.lockutils [req-76672f56-cbaf-44b6-8b80-61a1cedeeb2f req-cc36b601-5efb-4dcd-ac2c-677119a99b85 service nova] Lock "c96c4bcb-03f9-4961-aa17-8364aa4cf66e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.992485] env[61868]: DEBUG oslo_concurrency.lockutils [req-76672f56-cbaf-44b6-8b80-61a1cedeeb2f req-cc36b601-5efb-4dcd-ac2c-677119a99b85 service nova] Lock "c96c4bcb-03f9-4961-aa17-8364aa4cf66e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.992662] env[61868]: DEBUG nova.compute.manager [req-76672f56-cbaf-44b6-8b80-61a1cedeeb2f req-cc36b601-5efb-4dcd-ac2c-677119a99b85 service nova] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] No waiting events found dispatching network-vif-plugged-b0fdd6cb-31bd-4a5d-95cd-64d82af61df9 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1011.992765] env[61868]: WARNING nova.compute.manager [req-76672f56-cbaf-44b6-8b80-61a1cedeeb2f req-cc36b601-5efb-4dcd-ac2c-677119a99b85 service nova] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Received unexpected event network-vif-plugged-b0fdd6cb-31bd-4a5d-95cd-64d82af61df9 for instance with vm_state building and task_state spawning. [ 1012.052125] env[61868]: DEBUG oslo_concurrency.lockutils [None req-86c46a89-abde-41f6-8053-5e092f012a9a tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "9286cc76-1152-4fd4-9279-33d6ee4b9486" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.726s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.088454] env[61868]: DEBUG nova.network.neutron [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Successfully updated port: b0fdd6cb-31bd-4a5d-95cd-64d82af61df9 {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1012.171461] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dda1594-4054-4a63-ac8b-7824dad29018 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.179394] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-586fbd25-d919-49de-8f8c-c8d7cf919360 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.209909] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efae4d62-eac6-452b-8ab4-329338bb0032 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.217894] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c140ff69-22e7-4c84-8ed7-212e8f24568d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.231057] env[61868]: DEBUG nova.compute.provider_tree [None req-edf96371-ba85-4522-b84c-d581f28faebc tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1012.249997] env[61868]: DEBUG nova.virt.hardware [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1012.250239] env[61868]: DEBUG nova.virt.hardware [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1012.250406] env[61868]: DEBUG nova.virt.hardware [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1012.250597] env[61868]: DEBUG nova.virt.hardware [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1012.250746] env[61868]: DEBUG nova.virt.hardware [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1012.250895] env[61868]: DEBUG nova.virt.hardware [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1012.251107] env[61868]: DEBUG nova.virt.hardware [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1012.251273] env[61868]: DEBUG nova.virt.hardware [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1012.251442] env[61868]: DEBUG nova.virt.hardware [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1012.251607] env[61868]: DEBUG nova.virt.hardware [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1012.251782] env[61868]: DEBUG nova.virt.hardware [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1012.257076] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-54e945e0-86d5-4921-9db7-e76234eebec9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.272350] env[61868]: DEBUG oslo_vmware.api [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 1012.272350] env[61868]: value = "task-1315880" [ 1012.272350] env[61868]: _type = "Task" [ 1012.272350] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.279985] env[61868]: DEBUG oslo_vmware.api [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315880, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.590601] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "refresh_cache-c96c4bcb-03f9-4961-aa17-8364aa4cf66e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.590754] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquired lock "refresh_cache-c96c4bcb-03f9-4961-aa17-8364aa4cf66e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.590906] env[61868]: DEBUG nova.network.neutron [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1012.734092] env[61868]: DEBUG nova.scheduler.client.report [None req-edf96371-ba85-4522-b84c-d581f28faebc tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1012.782450] env[61868]: DEBUG oslo_vmware.api [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315880, 'name': ReconfigVM_Task, 'duration_secs': 0.186669} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.782769] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Updating instance '54dacd73-c760-49bd-9baf-ad2012a53264' progress to 33 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1013.121699] env[61868]: DEBUG nova.network.neutron [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1013.249785] env[61868]: DEBUG nova.network.neutron [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Updating instance_info_cache with network_info: [{"id": "b0fdd6cb-31bd-4a5d-95cd-64d82af61df9", "address": "fa:16:3e:cf:0d:98", "network": {"id": "f2e3f911-d826-43ad-911d-9f6985db7ee5", "bridge": "br-int", "label": "tempest-ImagesTestJSON-485234144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5658848e1b0e42929c04a4a8de40a291", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0fdd6cb-31", "ovs_interfaceid": "b0fdd6cb-31bd-4a5d-95cd-64d82af61df9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.288450] env[61868]: DEBUG nova.virt.hardware [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1013.288708] env[61868]: DEBUG nova.virt.hardware [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1013.288867] env[61868]: DEBUG nova.virt.hardware [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1013.289066] env[61868]: DEBUG nova.virt.hardware [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1013.289224] env[61868]: DEBUG nova.virt.hardware [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1013.289373] env[61868]: DEBUG nova.virt.hardware [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1013.289581] env[61868]: DEBUG nova.virt.hardware [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1013.289745] env[61868]: DEBUG nova.virt.hardware [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1013.289913] env[61868]: DEBUG nova.virt.hardware [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1013.290092] env[61868]: DEBUG nova.virt.hardware [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1013.290272] env[61868]: DEBUG nova.virt.hardware [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1013.295683] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Reconfiguring VM instance instance-0000005e to detach disk 2000 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1013.296200] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-13815cc6-748c-4c84-80df-e22489865f93 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.314290] env[61868]: DEBUG oslo_vmware.api [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 1013.314290] env[61868]: value = "task-1315881" [ 1013.314290] env[61868]: _type = "Task" [ 1013.314290] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.321952] env[61868]: DEBUG oslo_vmware.api [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315881, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.518531] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "d9e90641-6765-4bda-ab07-3506898f06a9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.518826] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "d9e90641-6765-4bda-ab07-3506898f06a9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.744064] env[61868]: DEBUG oslo_concurrency.lockutils [None req-edf96371-ba85-4522-b84c-d581f28faebc tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.222s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.751815] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Releasing lock "refresh_cache-c96c4bcb-03f9-4961-aa17-8364aa4cf66e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.752086] env[61868]: DEBUG nova.compute.manager [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Instance network_info: |[{"id": "b0fdd6cb-31bd-4a5d-95cd-64d82af61df9", "address": "fa:16:3e:cf:0d:98", "network": {"id": "f2e3f911-d826-43ad-911d-9f6985db7ee5", "bridge": "br-int", "label": "tempest-ImagesTestJSON-485234144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5658848e1b0e42929c04a4a8de40a291", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0fdd6cb-31", "ovs_interfaceid": "b0fdd6cb-31bd-4a5d-95cd-64d82af61df9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1013.752597] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cf:0d:98', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f1b507ed-cd2d-4c09-9d96-c47bde6a7774', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b0fdd6cb-31bd-4a5d-95cd-64d82af61df9', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1013.759884] env[61868]: DEBUG oslo.service.loopingcall [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1013.760119] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1013.760371] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-df83bce2-8bf3-4a1b-a403-baacf05aced1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.780755] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1013.780755] env[61868]: value = "task-1315882" [ 1013.780755] env[61868]: _type = "Task" [ 1013.780755] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.788426] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315882, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.822825] env[61868]: DEBUG oslo_vmware.api [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315881, 'name': ReconfigVM_Task, 'duration_secs': 0.225437} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.823100] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Reconfigured VM instance instance-0000005e to detach disk 2000 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1013.823847] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c484fa49-9f09-4a7f-9d20-2a3ec1b0b05c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.845340] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] 54dacd73-c760-49bd-9baf-ad2012a53264/54dacd73-c760-49bd-9baf-ad2012a53264.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1013.846020] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-46e8c1d6-9b28-44d6-8420-96ae4c3091e8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.863799] env[61868]: DEBUG oslo_vmware.api [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 1013.863799] env[61868]: value = "task-1315883" [ 1013.863799] env[61868]: _type = "Task" [ 1013.863799] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.872006] env[61868]: DEBUG oslo_vmware.api [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315883, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.021437] env[61868]: DEBUG nova.compute.manager [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1014.071762] env[61868]: DEBUG nova.compute.manager [req-272b81d6-7b90-48ee-9091-c6a37a46480b req-bb0401d0-bf89-4939-a0fc-6c57c2375158 service nova] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Received event network-changed-b0fdd6cb-31bd-4a5d-95cd-64d82af61df9 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1014.072129] env[61868]: DEBUG nova.compute.manager [req-272b81d6-7b90-48ee-9091-c6a37a46480b req-bb0401d0-bf89-4939-a0fc-6c57c2375158 service nova] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Refreshing instance network info cache due to event network-changed-b0fdd6cb-31bd-4a5d-95cd-64d82af61df9. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1014.072341] env[61868]: DEBUG oslo_concurrency.lockutils [req-272b81d6-7b90-48ee-9091-c6a37a46480b req-bb0401d0-bf89-4939-a0fc-6c57c2375158 service nova] Acquiring lock "refresh_cache-c96c4bcb-03f9-4961-aa17-8364aa4cf66e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.072548] env[61868]: DEBUG oslo_concurrency.lockutils [req-272b81d6-7b90-48ee-9091-c6a37a46480b req-bb0401d0-bf89-4939-a0fc-6c57c2375158 service nova] Acquired lock "refresh_cache-c96c4bcb-03f9-4961-aa17-8364aa4cf66e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.072776] env[61868]: DEBUG nova.network.neutron [req-272b81d6-7b90-48ee-9091-c6a37a46480b req-bb0401d0-bf89-4939-a0fc-6c57c2375158 service nova] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Refreshing network info cache for port b0fdd6cb-31bd-4a5d-95cd-64d82af61df9 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1014.292098] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315882, 'name': CreateVM_Task, 'duration_secs': 0.328931} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.292509] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1014.293336] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6732cc60-80ef-44f3-89ae-468b81ae48f9" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.293546] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6732cc60-80ef-44f3-89ae-468b81ae48f9" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.293980] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6732cc60-80ef-44f3-89ae-468b81ae48f9" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1014.294279] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-207b9515-bac9-4f4f-ae46-64dc43759019 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.301884] env[61868]: DEBUG oslo_vmware.api [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 1014.301884] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52183e24-9094-e91f-61ba-a7584a97b8bc" [ 1014.301884] env[61868]: _type = "Task" [ 1014.301884] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.307573] env[61868]: INFO nova.scheduler.client.report [None req-edf96371-ba85-4522-b84c-d581f28faebc tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Deleted allocation for migration 9f4d2d6e-1376-4d88-8eb6-b575d737488f [ 1014.318313] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6732cc60-80ef-44f3-89ae-468b81ae48f9" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.318593] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Processing image 6732cc60-80ef-44f3-89ae-468b81ae48f9 {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1014.318821] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6732cc60-80ef-44f3-89ae-468b81ae48f9/6732cc60-80ef-44f3-89ae-468b81ae48f9.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.318967] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6732cc60-80ef-44f3-89ae-468b81ae48f9/6732cc60-80ef-44f3-89ae-468b81ae48f9.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.319165] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1014.319423] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9388cbd4-b3fa-464a-a0c9-8b89fd3898e3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.327806] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1014.328035] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1014.328947] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-22bfb256-f6cc-4cea-956e-f454375c3c08 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.334231] env[61868]: DEBUG oslo_vmware.api [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 1014.334231] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52bd3c90-1fe2-ec09-5cdd-e33d25af91a6" [ 1014.334231] env[61868]: _type = "Task" [ 1014.334231] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.345170] env[61868]: DEBUG oslo_vmware.api [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52bd3c90-1fe2-ec09-5cdd-e33d25af91a6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.373283] env[61868]: DEBUG oslo_vmware.api [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315883, 'name': ReconfigVM_Task, 'duration_secs': 0.266428} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.373651] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Reconfigured VM instance instance-0000005e to attach disk [datastore2] 54dacd73-c760-49bd-9baf-ad2012a53264/54dacd73-c760-49bd-9baf-ad2012a53264.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1014.373962] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Updating instance '54dacd73-c760-49bd-9baf-ad2012a53264' progress to 50 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1014.544905] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.545219] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.546778] env[61868]: INFO nova.compute.claims [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1014.567051] env[61868]: DEBUG oslo_concurrency.lockutils [None req-cf962bea-f774-4ed5-8287-5701b9b2ef6d tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquiring lock "75fdbc11-1ef3-442d-bcf4-55069ba9ab62" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.567164] env[61868]: DEBUG oslo_concurrency.lockutils [None req-cf962bea-f774-4ed5-8287-5701b9b2ef6d tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lock "75fdbc11-1ef3-442d-bcf4-55069ba9ab62" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.567329] env[61868]: DEBUG oslo_concurrency.lockutils [None req-cf962bea-f774-4ed5-8287-5701b9b2ef6d tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquiring lock "75fdbc11-1ef3-442d-bcf4-55069ba9ab62-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.567518] env[61868]: DEBUG oslo_concurrency.lockutils [None req-cf962bea-f774-4ed5-8287-5701b9b2ef6d tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lock "75fdbc11-1ef3-442d-bcf4-55069ba9ab62-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.567690] env[61868]: DEBUG oslo_concurrency.lockutils [None req-cf962bea-f774-4ed5-8287-5701b9b2ef6d tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lock "75fdbc11-1ef3-442d-bcf4-55069ba9ab62-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.569638] env[61868]: INFO nova.compute.manager [None req-cf962bea-f774-4ed5-8287-5701b9b2ef6d tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Terminating instance [ 1014.571405] env[61868]: DEBUG nova.compute.manager [None req-cf962bea-f774-4ed5-8287-5701b9b2ef6d tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1014.571599] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-cf962bea-f774-4ed5-8287-5701b9b2ef6d tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1014.572478] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3910919f-77ca-40c2-b4de-fe42caa851fa {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.582176] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf962bea-f774-4ed5-8287-5701b9b2ef6d tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1014.582420] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-38fa1ace-e443-43a8-8882-9f385591ae81 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.589392] env[61868]: DEBUG oslo_vmware.api [None req-cf962bea-f774-4ed5-8287-5701b9b2ef6d tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the task: (returnval){ [ 1014.589392] env[61868]: value = "task-1315884" [ 1014.589392] env[61868]: _type = "Task" [ 1014.589392] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.597368] env[61868]: DEBUG oslo_vmware.api [None req-cf962bea-f774-4ed5-8287-5701b9b2ef6d tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315884, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.780675] env[61868]: DEBUG nova.network.neutron [req-272b81d6-7b90-48ee-9091-c6a37a46480b req-bb0401d0-bf89-4939-a0fc-6c57c2375158 service nova] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Updated VIF entry in instance network info cache for port b0fdd6cb-31bd-4a5d-95cd-64d82af61df9. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1014.781049] env[61868]: DEBUG nova.network.neutron [req-272b81d6-7b90-48ee-9091-c6a37a46480b req-bb0401d0-bf89-4939-a0fc-6c57c2375158 service nova] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Updating instance_info_cache with network_info: [{"id": "b0fdd6cb-31bd-4a5d-95cd-64d82af61df9", "address": "fa:16:3e:cf:0d:98", "network": {"id": "f2e3f911-d826-43ad-911d-9f6985db7ee5", "bridge": "br-int", "label": "tempest-ImagesTestJSON-485234144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5658848e1b0e42929c04a4a8de40a291", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0fdd6cb-31", "ovs_interfaceid": "b0fdd6cb-31bd-4a5d-95cd-64d82af61df9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.815214] env[61868]: DEBUG oslo_concurrency.lockutils [None req-edf96371-ba85-4522-b84c-d581f28faebc tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "5d2b56d5-a4e0-4dba-9158-71036d60a40b" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 8.281s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.844568] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Preparing fetch location {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1014.844768] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Fetch image to [datastore2] OSTACK_IMG_d3e07e05-acdb-45f2-bf23-08f368a227b5/OSTACK_IMG_d3e07e05-acdb-45f2-bf23-08f368a227b5.vmdk {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1014.844957] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Downloading stream optimized image 6732cc60-80ef-44f3-89ae-468b81ae48f9 to [datastore2] OSTACK_IMG_d3e07e05-acdb-45f2-bf23-08f368a227b5/OSTACK_IMG_d3e07e05-acdb-45f2-bf23-08f368a227b5.vmdk on the data store datastore2 as vApp {{(pid=61868) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1014.845162] env[61868]: DEBUG nova.virt.vmwareapi.images [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Downloading image file data 6732cc60-80ef-44f3-89ae-468b81ae48f9 to the ESX as VM named 'OSTACK_IMG_d3e07e05-acdb-45f2-bf23-08f368a227b5' {{(pid=61868) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1014.882258] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2081c690-b87b-49a9-8a57-7ebe8e2fc240 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.899704] env[61868]: DEBUG oslo_concurrency.lockutils [None req-09663916-c9cf-4eeb-8d04-a9f10db6150d tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "5d2b56d5-a4e0-4dba-9158-71036d60a40b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.899950] env[61868]: DEBUG oslo_concurrency.lockutils [None req-09663916-c9cf-4eeb-8d04-a9f10db6150d tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "5d2b56d5-a4e0-4dba-9158-71036d60a40b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.900165] env[61868]: DEBUG oslo_concurrency.lockutils [None req-09663916-c9cf-4eeb-8d04-a9f10db6150d tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "5d2b56d5-a4e0-4dba-9158-71036d60a40b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.900351] env[61868]: DEBUG oslo_concurrency.lockutils [None req-09663916-c9cf-4eeb-8d04-a9f10db6150d tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "5d2b56d5-a4e0-4dba-9158-71036d60a40b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.900522] env[61868]: DEBUG oslo_concurrency.lockutils [None req-09663916-c9cf-4eeb-8d04-a9f10db6150d tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "5d2b56d5-a4e0-4dba-9158-71036d60a40b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.904290] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9567692-6854-4ce1-9ab0-7ded7242291c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.907247] env[61868]: INFO nova.compute.manager [None req-09663916-c9cf-4eeb-8d04-a9f10db6150d tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Terminating instance [ 1014.909040] env[61868]: DEBUG nova.compute.manager [None req-09663916-c9cf-4eeb-8d04-a9f10db6150d tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1014.909241] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-09663916-c9cf-4eeb-8d04-a9f10db6150d tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1014.910268] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbfbafcb-8209-4bfa-a0e0-1a583d38e9ba {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.928913] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Updating instance '54dacd73-c760-49bd-9baf-ad2012a53264' progress to 67 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1014.934343] env[61868]: DEBUG oslo_vmware.rw_handles [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1014.934343] env[61868]: value = "resgroup-9" [ 1014.934343] env[61868]: _type = "ResourcePool" [ 1014.934343] env[61868]: }. {{(pid=61868) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1014.935082] env[61868]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-010a895d-23c8-4a1b-9168-f43606f69261 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.952238] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-09663916-c9cf-4eeb-8d04-a9f10db6150d tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1014.952907] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-95b42586-1f8f-47d7-b51c-67c460012381 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.956844] env[61868]: DEBUG oslo_vmware.rw_handles [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lease: (returnval){ [ 1014.956844] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52b4aa70-7691-56be-05a3-83a451089a60" [ 1014.956844] env[61868]: _type = "HttpNfcLease" [ 1014.956844] env[61868]: } obtained for vApp import into resource pool (val){ [ 1014.956844] env[61868]: value = "resgroup-9" [ 1014.956844] env[61868]: _type = "ResourcePool" [ 1014.956844] env[61868]: }. {{(pid=61868) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1014.957167] env[61868]: DEBUG oslo_vmware.api [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the lease: (returnval){ [ 1014.957167] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52b4aa70-7691-56be-05a3-83a451089a60" [ 1014.957167] env[61868]: _type = "HttpNfcLease" [ 1014.957167] env[61868]: } to be ready. {{(pid=61868) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1014.961435] env[61868]: DEBUG oslo_vmware.api [None req-09663916-c9cf-4eeb-8d04-a9f10db6150d tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 1014.961435] env[61868]: value = "task-1315886" [ 1014.961435] env[61868]: _type = "Task" [ 1014.961435] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.965814] env[61868]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1014.965814] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52b4aa70-7691-56be-05a3-83a451089a60" [ 1014.965814] env[61868]: _type = "HttpNfcLease" [ 1014.965814] env[61868]: } is initializing. {{(pid=61868) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1014.971574] env[61868]: DEBUG oslo_vmware.api [None req-09663916-c9cf-4eeb-8d04-a9f10db6150d tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315886, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.100385] env[61868]: DEBUG oslo_vmware.api [None req-cf962bea-f774-4ed5-8287-5701b9b2ef6d tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315884, 'name': PowerOffVM_Task, 'duration_secs': 0.179033} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.100709] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf962bea-f774-4ed5-8287-5701b9b2ef6d tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1015.100903] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-cf962bea-f774-4ed5-8287-5701b9b2ef6d tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1015.101193] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a0fa4ed3-4faa-4dd8-90c9-89df9c84de8c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.155277] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-cf962bea-f774-4ed5-8287-5701b9b2ef6d tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1015.155595] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-cf962bea-f774-4ed5-8287-5701b9b2ef6d tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1015.155827] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf962bea-f774-4ed5-8287-5701b9b2ef6d tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Deleting the datastore file [datastore1] 75fdbc11-1ef3-442d-bcf4-55069ba9ab62 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1015.156132] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6443e7e7-cc3e-45ed-bdc6-f5b192fc1bdc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.162337] env[61868]: DEBUG oslo_vmware.api [None req-cf962bea-f774-4ed5-8287-5701b9b2ef6d tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for the task: (returnval){ [ 1015.162337] env[61868]: value = "task-1315888" [ 1015.162337] env[61868]: _type = "Task" [ 1015.162337] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.170722] env[61868]: DEBUG oslo_vmware.api [None req-cf962bea-f774-4ed5-8287-5701b9b2ef6d tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315888, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.285119] env[61868]: DEBUG oslo_concurrency.lockutils [req-272b81d6-7b90-48ee-9091-c6a37a46480b req-bb0401d0-bf89-4939-a0fc-6c57c2375158 service nova] Releasing lock "refresh_cache-c96c4bcb-03f9-4961-aa17-8364aa4cf66e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.292597] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "0316bfec-128e-41dc-ad3a-9ba3844b95fa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.292937] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "0316bfec-128e-41dc-ad3a-9ba3844b95fa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.467062] env[61868]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1015.467062] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52b4aa70-7691-56be-05a3-83a451089a60" [ 1015.467062] env[61868]: _type = "HttpNfcLease" [ 1015.467062] env[61868]: } is initializing. {{(pid=61868) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1015.472276] env[61868]: DEBUG oslo_vmware.api [None req-09663916-c9cf-4eeb-8d04-a9f10db6150d tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315886, 'name': PowerOffVM_Task, 'duration_secs': 0.18661} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.472582] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-09663916-c9cf-4eeb-8d04-a9f10db6150d tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1015.472761] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-09663916-c9cf-4eeb-8d04-a9f10db6150d tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1015.473010] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cd3e07ab-060d-43ef-a884-5ad7929f390e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.487152] env[61868]: DEBUG nova.network.neutron [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Port cbd81a07-4122-41bd-bb91-d47b6bc3918d binding to destination host cpu-1 is already ACTIVE {{(pid=61868) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1015.532056] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-09663916-c9cf-4eeb-8d04-a9f10db6150d tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1015.532248] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-09663916-c9cf-4eeb-8d04-a9f10db6150d tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1015.532482] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-09663916-c9cf-4eeb-8d04-a9f10db6150d tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Deleting the datastore file [datastore1] 5d2b56d5-a4e0-4dba-9158-71036d60a40b {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1015.533121] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d17ece94-f1db-4121-9801-ec9b8b2d351c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.539425] env[61868]: DEBUG oslo_vmware.api [None req-09663916-c9cf-4eeb-8d04-a9f10db6150d tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 1015.539425] env[61868]: value = "task-1315890" [ 1015.539425] env[61868]: _type = "Task" [ 1015.539425] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.547306] env[61868]: DEBUG oslo_vmware.api [None req-09663916-c9cf-4eeb-8d04-a9f10db6150d tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315890, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.672266] env[61868]: DEBUG oslo_vmware.api [None req-cf962bea-f774-4ed5-8287-5701b9b2ef6d tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Task: {'id': task-1315888, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134884} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.672571] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf962bea-f774-4ed5-8287-5701b9b2ef6d tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1015.672788] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-cf962bea-f774-4ed5-8287-5701b9b2ef6d tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1015.672994] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-cf962bea-f774-4ed5-8287-5701b9b2ef6d tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1015.673202] env[61868]: INFO nova.compute.manager [None req-cf962bea-f774-4ed5-8287-5701b9b2ef6d tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1015.673447] env[61868]: DEBUG oslo.service.loopingcall [None req-cf962bea-f774-4ed5-8287-5701b9b2ef6d tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1015.673645] env[61868]: DEBUG nova.compute.manager [-] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1015.673761] env[61868]: DEBUG nova.network.neutron [-] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1015.700841] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d6dfe2e-19a3-4991-af0b-da082202f44c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.709148] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5cc8d05-acf8-43ba-a6d4-e8460ba1c497 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.743392] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dda38f73-0ae1-4049-ac11-bbcc7617fdc0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.751273] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-980aa123-8612-43b3-8c59-942a5c5f08fd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.767112] env[61868]: DEBUG nova.compute.provider_tree [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1015.795496] env[61868]: DEBUG nova.compute.manager [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1015.966980] env[61868]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1015.966980] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52b4aa70-7691-56be-05a3-83a451089a60" [ 1015.966980] env[61868]: _type = "HttpNfcLease" [ 1015.966980] env[61868]: } is ready. {{(pid=61868) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1015.966980] env[61868]: DEBUG oslo_vmware.rw_handles [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1015.966980] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52b4aa70-7691-56be-05a3-83a451089a60" [ 1015.966980] env[61868]: _type = "HttpNfcLease" [ 1015.966980] env[61868]: }. {{(pid=61868) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1015.967492] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-351c0402-5181-40cf-85fd-38d7746942f2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.973947] env[61868]: DEBUG oslo_vmware.rw_handles [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a8ea11-d84a-e4e3-4201-160386d52aba/disk-0.vmdk from lease info. {{(pid=61868) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1015.974139] env[61868]: DEBUG oslo_vmware.rw_handles [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a8ea11-d84a-e4e3-4201-160386d52aba/disk-0.vmdk. {{(pid=61868) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1016.039675] env[61868]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b2ed7e52-0ab6-42f1-8271-099da19dec7c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.050780] env[61868]: DEBUG oslo_vmware.api [None req-09663916-c9cf-4eeb-8d04-a9f10db6150d tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315890, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.169754} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.051823] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-09663916-c9cf-4eeb-8d04-a9f10db6150d tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1016.052033] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-09663916-c9cf-4eeb-8d04-a9f10db6150d tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1016.052220] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-09663916-c9cf-4eeb-8d04-a9f10db6150d tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1016.052424] env[61868]: INFO nova.compute.manager [None req-09663916-c9cf-4eeb-8d04-a9f10db6150d tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1016.052662] env[61868]: DEBUG oslo.service.loopingcall [None req-09663916-c9cf-4eeb-8d04-a9f10db6150d tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1016.052869] env[61868]: DEBUG nova.compute.manager [-] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1016.052965] env[61868]: DEBUG nova.network.neutron [-] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1016.100159] env[61868]: DEBUG nova.compute.manager [req-1b34adec-6941-41c5-83c8-206100720582 req-2959785f-aca9-4250-a18c-e5398adba23c service nova] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Received event network-vif-deleted-21d6cab8-f691-425c-bfa8-9d2f1c294b64 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1016.100380] env[61868]: INFO nova.compute.manager [req-1b34adec-6941-41c5-83c8-206100720582 req-2959785f-aca9-4250-a18c-e5398adba23c service nova] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Neutron deleted interface 21d6cab8-f691-425c-bfa8-9d2f1c294b64; detaching it from the instance and deleting it from the info cache [ 1016.100556] env[61868]: DEBUG nova.network.neutron [req-1b34adec-6941-41c5-83c8-206100720582 req-2959785f-aca9-4250-a18c-e5398adba23c service nova] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.270131] env[61868]: DEBUG nova.scheduler.client.report [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1016.319055] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.423783] env[61868]: DEBUG nova.network.neutron [-] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.552715] env[61868]: DEBUG oslo_concurrency.lockutils [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "54dacd73-c760-49bd-9baf-ad2012a53264-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.553015] env[61868]: DEBUG oslo_concurrency.lockutils [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "54dacd73-c760-49bd-9baf-ad2012a53264-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.553209] env[61868]: DEBUG oslo_concurrency.lockutils [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "54dacd73-c760-49bd-9baf-ad2012a53264-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.603641] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f975d814-6ffa-4738-bbcf-8f5b7624385f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.615914] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2f74f77-e576-490e-afd9-bfc1fdf88e73 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.649207] env[61868]: DEBUG nova.compute.manager [req-1b34adec-6941-41c5-83c8-206100720582 req-2959785f-aca9-4250-a18c-e5398adba23c service nova] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Detach interface failed, port_id=21d6cab8-f691-425c-bfa8-9d2f1c294b64, reason: Instance 75fdbc11-1ef3-442d-bcf4-55069ba9ab62 could not be found. {{(pid=61868) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1016.776454] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.231s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.777195] env[61868]: DEBUG nova.compute.manager [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1016.780758] env[61868]: DEBUG oslo_vmware.rw_handles [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Completed reading data from the image iterator. {{(pid=61868) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1016.780958] env[61868]: DEBUG oslo_vmware.rw_handles [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a8ea11-d84a-e4e3-4201-160386d52aba/disk-0.vmdk. {{(pid=61868) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1016.781314] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.462s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.783008] env[61868]: INFO nova.compute.claims [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1016.785553] env[61868]: DEBUG nova.network.neutron [-] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.787651] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bae1d8d-658c-4c15-9077-59a980204a1f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.795274] env[61868]: DEBUG oslo_vmware.rw_handles [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a8ea11-d84a-e4e3-4201-160386d52aba/disk-0.vmdk is in state: ready. {{(pid=61868) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1016.795487] env[61868]: DEBUG oslo_vmware.rw_handles [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a8ea11-d84a-e4e3-4201-160386d52aba/disk-0.vmdk. {{(pid=61868) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1016.795728] env[61868]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-0d1ab13c-4883-4e80-8c1b-33e86b405afb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.927776] env[61868]: INFO nova.compute.manager [-] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Took 1.25 seconds to deallocate network for instance. [ 1016.978596] env[61868]: DEBUG oslo_vmware.rw_handles [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a8ea11-d84a-e4e3-4201-160386d52aba/disk-0.vmdk. {{(pid=61868) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1016.978851] env[61868]: INFO nova.virt.vmwareapi.images [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Downloaded image file data 6732cc60-80ef-44f3-89ae-468b81ae48f9 [ 1016.979911] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c7df97a-8f88-41e8-a52c-1d39427a1c65 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.995450] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9ee24bc8-b24c-4f61-83eb-3b74e8098f5a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.019185] env[61868]: INFO nova.virt.vmwareapi.images [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] The imported VM was unregistered [ 1017.021672] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Caching image {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1017.021906] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Creating directory with path [datastore2] devstack-image-cache_base/6732cc60-80ef-44f3-89ae-468b81ae48f9 {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1017.022333] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-57a62f87-1b31-44ec-bff8-1e22a80f3c82 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.032499] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Created directory with path [datastore2] devstack-image-cache_base/6732cc60-80ef-44f3-89ae-468b81ae48f9 {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1017.032710] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_d3e07e05-acdb-45f2-bf23-08f368a227b5/OSTACK_IMG_d3e07e05-acdb-45f2-bf23-08f368a227b5.vmdk to [datastore2] devstack-image-cache_base/6732cc60-80ef-44f3-89ae-468b81ae48f9/6732cc60-80ef-44f3-89ae-468b81ae48f9.vmdk. {{(pid=61868) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1017.032967] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-01486372-cc70-45b0-a3b3-b28f6500a24d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.041439] env[61868]: DEBUG oslo_vmware.api [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 1017.041439] env[61868]: value = "task-1315892" [ 1017.041439] env[61868]: _type = "Task" [ 1017.041439] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.048640] env[61868]: DEBUG oslo_vmware.api [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315892, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.297969] env[61868]: DEBUG nova.compute.utils [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1017.297969] env[61868]: INFO nova.compute.manager [-] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Took 1.24 seconds to deallocate network for instance. [ 1017.297969] env[61868]: DEBUG nova.compute.manager [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1017.297969] env[61868]: DEBUG nova.network.neutron [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1017.345146] env[61868]: DEBUG nova.policy [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'daad105e7edf4fb0ae0b2e685bfd92e9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b520c3ac58074e8d9b0bfafb817244a8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 1017.433970] env[61868]: DEBUG oslo_concurrency.lockutils [None req-cf962bea-f774-4ed5-8287-5701b9b2ef6d tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.556256] env[61868]: DEBUG oslo_vmware.api [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315892, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.599298] env[61868]: DEBUG oslo_concurrency.lockutils [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "refresh_cache-54dacd73-c760-49bd-9baf-ad2012a53264" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.599298] env[61868]: DEBUG oslo_concurrency.lockutils [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquired lock "refresh_cache-54dacd73-c760-49bd-9baf-ad2012a53264" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.599298] env[61868]: DEBUG nova.network.neutron [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1017.618257] env[61868]: DEBUG nova.network.neutron [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Successfully created port: 30ef9d89-a3e3-4d38-8a41-41a35000db0d {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1017.801362] env[61868]: DEBUG nova.compute.manager [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1017.806848] env[61868]: DEBUG oslo_concurrency.lockutils [None req-09663916-c9cf-4eeb-8d04-a9f10db6150d tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.987829] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d57894ff-964f-4511-9312-6da915db060a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.997600] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e16c0b2-5e61-49ce-9eec-dcc47ef8d29a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.031102] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fec52085-4fe0-4372-803c-298503124c2b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.041247] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65acfcf1-66c7-4eb5-85d0-a85f14a45c02 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.055601] env[61868]: DEBUG oslo_vmware.api [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315892, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.063584] env[61868]: DEBUG nova.compute.provider_tree [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1018.128441] env[61868]: DEBUG nova.compute.manager [req-12a4af08-9727-42ee-baf1-687b87999db0 req-6dec57ef-94ea-44da-8224-077d36c22867 service nova] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Received event network-vif-deleted-9023a9b8-ad2a-42b6-b2cd-b2cdb86dd9cb {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1018.425354] env[61868]: DEBUG nova.network.neutron [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Updating instance_info_cache with network_info: [{"id": "cbd81a07-4122-41bd-bb91-d47b6bc3918d", "address": "fa:16:3e:31:4b:86", "network": {"id": "5b4c5906-b873-4369-855b-e0898ad84812", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-254512637-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "812e115cc1114d0e8536924c4f5556f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcbd81a07-41", "ovs_interfaceid": "cbd81a07-4122-41bd-bb91-d47b6bc3918d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1018.555973] env[61868]: DEBUG oslo_vmware.api [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315892, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.568448] env[61868]: DEBUG nova.scheduler.client.report [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1018.810658] env[61868]: DEBUG nova.compute.manager [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1018.838197] env[61868]: DEBUG nova.virt.hardware [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1018.838495] env[61868]: DEBUG nova.virt.hardware [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1018.838716] env[61868]: DEBUG nova.virt.hardware [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1018.838964] env[61868]: DEBUG nova.virt.hardware [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1018.839216] env[61868]: DEBUG nova.virt.hardware [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1018.839403] env[61868]: DEBUG nova.virt.hardware [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1018.839637] env[61868]: DEBUG nova.virt.hardware [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1018.839866] env[61868]: DEBUG nova.virt.hardware [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1018.840074] env[61868]: DEBUG nova.virt.hardware [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1018.840250] env[61868]: DEBUG nova.virt.hardware [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1018.840469] env[61868]: DEBUG nova.virt.hardware [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1018.841461] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f51c15f-01f7-43b4-a3af-ee5b0e2b5386 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.852018] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69859fce-67a7-4b1e-900a-23e42205a63a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.928326] env[61868]: DEBUG oslo_concurrency.lockutils [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Releasing lock "refresh_cache-54dacd73-c760-49bd-9baf-ad2012a53264" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.057237] env[61868]: DEBUG oslo_vmware.api [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315892, 'name': MoveVirtualDisk_Task} progress is 88%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.072951] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "interface-a96cab5a-c053-4e45-96f6-2aba0a819110-0bbbc97a-2c41-4d5d-a093-fca57a1d543b" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.073274] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "interface-a96cab5a-c053-4e45-96f6-2aba0a819110-0bbbc97a-2c41-4d5d-a093-fca57a1d543b" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.073755] env[61868]: DEBUG nova.objects.instance [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lazy-loading 'flavor' on Instance uuid a96cab5a-c053-4e45-96f6-2aba0a819110 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1019.075572] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.294s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.076170] env[61868]: DEBUG nova.compute.manager [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1019.078883] env[61868]: DEBUG oslo_concurrency.lockutils [None req-cf962bea-f774-4ed5-8287-5701b9b2ef6d tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.645s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.079188] env[61868]: DEBUG nova.objects.instance [None req-cf962bea-f774-4ed5-8287-5701b9b2ef6d tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lazy-loading 'resources' on Instance uuid 75fdbc11-1ef3-442d-bcf4-55069ba9ab62 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1019.163167] env[61868]: DEBUG nova.network.neutron [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Successfully updated port: 30ef9d89-a3e3-4d38-8a41-41a35000db0d {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1019.456040] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4f7fc71-0fdd-4554-a2a5-cd2a09421b9b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.475742] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36f1b6f2-ca30-467e-a1be-6c0d39ff5eaa {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.482584] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Updating instance '54dacd73-c760-49bd-9baf-ad2012a53264' progress to 83 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1019.555852] env[61868]: DEBUG oslo_vmware.api [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315892, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.340209} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.556209] env[61868]: INFO nova.virt.vmwareapi.ds_util [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_d3e07e05-acdb-45f2-bf23-08f368a227b5/OSTACK_IMG_d3e07e05-acdb-45f2-bf23-08f368a227b5.vmdk to [datastore2] devstack-image-cache_base/6732cc60-80ef-44f3-89ae-468b81ae48f9/6732cc60-80ef-44f3-89ae-468b81ae48f9.vmdk. [ 1019.556408] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Cleaning up location [datastore2] OSTACK_IMG_d3e07e05-acdb-45f2-bf23-08f368a227b5 {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1019.556578] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_d3e07e05-acdb-45f2-bf23-08f368a227b5 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1019.556831] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-360c8119-a9aa-4e04-927b-7d9c402fa405 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.563560] env[61868]: DEBUG oslo_vmware.api [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 1019.563560] env[61868]: value = "task-1315893" [ 1019.563560] env[61868]: _type = "Task" [ 1019.563560] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.570886] env[61868]: DEBUG oslo_vmware.api [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315893, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.582769] env[61868]: DEBUG nova.compute.utils [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1019.588025] env[61868]: DEBUG nova.compute.manager [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1019.588025] env[61868]: DEBUG nova.network.neutron [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1019.641157] env[61868]: DEBUG nova.policy [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e02ffdc140274733a081fd53c4acc202', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a60c3c20950e4156b64c3b4c61b9f0f2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 1019.665735] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "refresh_cache-d9e90641-6765-4bda-ab07-3506898f06a9" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.665900] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquired lock "refresh_cache-d9e90641-6765-4bda-ab07-3506898f06a9" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.666091] env[61868]: DEBUG nova.network.neutron [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1019.681742] env[61868]: DEBUG nova.objects.instance [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lazy-loading 'pci_requests' on Instance uuid a96cab5a-c053-4e45-96f6-2aba0a819110 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1019.739918] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2cd1614-7d92-4f74-96ce-e3682f0f58c2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.748048] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71ee7f48-43e3-4599-8ae0-73c5f8be771f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.778995] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24610617-a67c-4cd9-afad-8643cb83cfee {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.786255] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9a97ad1-54db-4c26-a46c-0e58fd07162a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.799017] env[61868]: DEBUG nova.compute.provider_tree [None req-cf962bea-f774-4ed5-8287-5701b9b2ef6d tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1019.988584] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1019.988875] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-533b6caa-be49-4379-a35f-7ac3e7c850f0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.997044] env[61868]: DEBUG oslo_vmware.api [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 1019.997044] env[61868]: value = "task-1315894" [ 1019.997044] env[61868]: _type = "Task" [ 1019.997044] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.004938] env[61868]: DEBUG oslo_vmware.api [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315894, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.075860] env[61868]: DEBUG oslo_vmware.api [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315893, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.035565} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.076228] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1020.076515] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6732cc60-80ef-44f3-89ae-468b81ae48f9/6732cc60-80ef-44f3-89ae-468b81ae48f9.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.076893] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6732cc60-80ef-44f3-89ae-468b81ae48f9/6732cc60-80ef-44f3-89ae-468b81ae48f9.vmdk to [datastore2] c96c4bcb-03f9-4961-aa17-8364aa4cf66e/c96c4bcb-03f9-4961-aa17-8364aa4cf66e.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1020.077243] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4376870e-111e-4033-9517-2cece2b8bca4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.084761] env[61868]: DEBUG oslo_vmware.api [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 1020.084761] env[61868]: value = "task-1315895" [ 1020.084761] env[61868]: _type = "Task" [ 1020.084761] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.089465] env[61868]: DEBUG nova.compute.manager [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1020.100332] env[61868]: DEBUG oslo_vmware.api [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315895, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.101399] env[61868]: DEBUG nova.network.neutron [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Successfully created port: 56a4bffb-5418-4c4c-a6b6-20edaf0e3782 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1020.186290] env[61868]: DEBUG nova.objects.base [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61868) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1020.186538] env[61868]: DEBUG nova.network.neutron [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1020.212160] env[61868]: DEBUG nova.network.neutron [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1020.246435] env[61868]: DEBUG nova.compute.manager [req-fcc183b8-4341-4d68-a3e5-291770111d12 req-836297f3-ca52-432a-ba94-416b25785b7f service nova] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Received event network-vif-plugged-30ef9d89-a3e3-4d38-8a41-41a35000db0d {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1020.246699] env[61868]: DEBUG oslo_concurrency.lockutils [req-fcc183b8-4341-4d68-a3e5-291770111d12 req-836297f3-ca52-432a-ba94-416b25785b7f service nova] Acquiring lock "d9e90641-6765-4bda-ab07-3506898f06a9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.246912] env[61868]: DEBUG oslo_concurrency.lockutils [req-fcc183b8-4341-4d68-a3e5-291770111d12 req-836297f3-ca52-432a-ba94-416b25785b7f service nova] Lock "d9e90641-6765-4bda-ab07-3506898f06a9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.247510] env[61868]: DEBUG oslo_concurrency.lockutils [req-fcc183b8-4341-4d68-a3e5-291770111d12 req-836297f3-ca52-432a-ba94-416b25785b7f service nova] Lock "d9e90641-6765-4bda-ab07-3506898f06a9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.247510] env[61868]: DEBUG nova.compute.manager [req-fcc183b8-4341-4d68-a3e5-291770111d12 req-836297f3-ca52-432a-ba94-416b25785b7f service nova] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] No waiting events found dispatching network-vif-plugged-30ef9d89-a3e3-4d38-8a41-41a35000db0d {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1020.247510] env[61868]: WARNING nova.compute.manager [req-fcc183b8-4341-4d68-a3e5-291770111d12 req-836297f3-ca52-432a-ba94-416b25785b7f service nova] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Received unexpected event network-vif-plugged-30ef9d89-a3e3-4d38-8a41-41a35000db0d for instance with vm_state building and task_state spawning. [ 1020.247714] env[61868]: DEBUG nova.compute.manager [req-fcc183b8-4341-4d68-a3e5-291770111d12 req-836297f3-ca52-432a-ba94-416b25785b7f service nova] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Received event network-changed-30ef9d89-a3e3-4d38-8a41-41a35000db0d {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1020.247763] env[61868]: DEBUG nova.compute.manager [req-fcc183b8-4341-4d68-a3e5-291770111d12 req-836297f3-ca52-432a-ba94-416b25785b7f service nova] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Refreshing instance network info cache due to event network-changed-30ef9d89-a3e3-4d38-8a41-41a35000db0d. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1020.248017] env[61868]: DEBUG oslo_concurrency.lockutils [req-fcc183b8-4341-4d68-a3e5-291770111d12 req-836297f3-ca52-432a-ba94-416b25785b7f service nova] Acquiring lock "refresh_cache-d9e90641-6765-4bda-ab07-3506898f06a9" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.296491] env[61868]: DEBUG nova.policy [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c14b4ff01717495ca97c7f35f91c2995', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd7fc76299baf4a97b57139e5f1caa16e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 1020.301769] env[61868]: DEBUG nova.scheduler.client.report [None req-cf962bea-f774-4ed5-8287-5701b9b2ef6d tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1020.485263] env[61868]: DEBUG nova.network.neutron [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Updating instance_info_cache with network_info: [{"id": "30ef9d89-a3e3-4d38-8a41-41a35000db0d", "address": "fa:16:3e:33:d2:c4", "network": {"id": "77f33b52-e4eb-4b72-85d4-5991a2b5c36d", "bridge": "br-int", "label": "tempest-ServersTestJSON-748635995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b520c3ac58074e8d9b0bfafb817244a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30ef9d89-a3", "ovs_interfaceid": "30ef9d89-a3e3-4d38-8a41-41a35000db0d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.515504] env[61868]: DEBUG oslo_vmware.api [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315894, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.602988] env[61868]: DEBUG oslo_vmware.api [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315895, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.807555] env[61868]: DEBUG oslo_concurrency.lockutils [None req-cf962bea-f774-4ed5-8287-5701b9b2ef6d tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.729s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.810381] env[61868]: DEBUG oslo_concurrency.lockutils [None req-09663916-c9cf-4eeb-8d04-a9f10db6150d tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.004s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.810950] env[61868]: DEBUG oslo_concurrency.lockutils [None req-09663916-c9cf-4eeb-8d04-a9f10db6150d tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.836270] env[61868]: INFO nova.scheduler.client.report [None req-cf962bea-f774-4ed5-8287-5701b9b2ef6d tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Deleted allocations for instance 75fdbc11-1ef3-442d-bcf4-55069ba9ab62 [ 1020.838798] env[61868]: INFO nova.scheduler.client.report [None req-09663916-c9cf-4eeb-8d04-a9f10db6150d tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Deleted allocations for instance 5d2b56d5-a4e0-4dba-9158-71036d60a40b [ 1020.988202] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Releasing lock "refresh_cache-d9e90641-6765-4bda-ab07-3506898f06a9" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.988705] env[61868]: DEBUG nova.compute.manager [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Instance network_info: |[{"id": "30ef9d89-a3e3-4d38-8a41-41a35000db0d", "address": "fa:16:3e:33:d2:c4", "network": {"id": "77f33b52-e4eb-4b72-85d4-5991a2b5c36d", "bridge": "br-int", "label": "tempest-ServersTestJSON-748635995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b520c3ac58074e8d9b0bfafb817244a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30ef9d89-a3", "ovs_interfaceid": "30ef9d89-a3e3-4d38-8a41-41a35000db0d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1020.989249] env[61868]: DEBUG oslo_concurrency.lockutils [req-fcc183b8-4341-4d68-a3e5-291770111d12 req-836297f3-ca52-432a-ba94-416b25785b7f service nova] Acquired lock "refresh_cache-d9e90641-6765-4bda-ab07-3506898f06a9" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.989894] env[61868]: DEBUG nova.network.neutron [req-fcc183b8-4341-4d68-a3e5-291770111d12 req-836297f3-ca52-432a-ba94-416b25785b7f service nova] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Refreshing network info cache for port 30ef9d89-a3e3-4d38-8a41-41a35000db0d {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1020.991414] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:33:d2:c4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56136ef6-99d7-4562-9a9f-d66fec951c5c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '30ef9d89-a3e3-4d38-8a41-41a35000db0d', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1021.005570] env[61868]: DEBUG oslo.service.loopingcall [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1021.007282] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1021.012685] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3ee619d7-c669-40bc-a545-c9fe01dc20e0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.044780] env[61868]: DEBUG oslo_vmware.api [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315894, 'name': PowerOnVM_Task, 'duration_secs': 0.562129} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.046326] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1021.046536] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-650038c9-8be3-4359-b3bd-6de90e995e61 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Updating instance '54dacd73-c760-49bd-9baf-ad2012a53264' progress to 100 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1021.050052] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1021.050052] env[61868]: value = "task-1315896" [ 1021.050052] env[61868]: _type = "Task" [ 1021.050052] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.062154] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315896, 'name': CreateVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.096317] env[61868]: DEBUG oslo_vmware.api [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315895, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.105396] env[61868]: DEBUG nova.compute.manager [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1021.140952] env[61868]: DEBUG nova.virt.hardware [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1021.141244] env[61868]: DEBUG nova.virt.hardware [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1021.141434] env[61868]: DEBUG nova.virt.hardware [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1021.141647] env[61868]: DEBUG nova.virt.hardware [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1021.141800] env[61868]: DEBUG nova.virt.hardware [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1021.141954] env[61868]: DEBUG nova.virt.hardware [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1021.142224] env[61868]: DEBUG nova.virt.hardware [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1021.142396] env[61868]: DEBUG nova.virt.hardware [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1021.142606] env[61868]: DEBUG nova.virt.hardware [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1021.142775] env[61868]: DEBUG nova.virt.hardware [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1021.142969] env[61868]: DEBUG nova.virt.hardware [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1021.143901] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45730463-a334-4a6b-915d-21fd6432f447 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.152957] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-538a493a-d72c-4245-bfb4-84e60dc9cdd4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.350812] env[61868]: DEBUG oslo_concurrency.lockutils [None req-cf962bea-f774-4ed5-8287-5701b9b2ef6d tempest-ServersNegativeTestJSON-730537016 tempest-ServersNegativeTestJSON-730537016-project-member] Lock "75fdbc11-1ef3-442d-bcf4-55069ba9ab62" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.783s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.352008] env[61868]: DEBUG oslo_concurrency.lockutils [None req-09663916-c9cf-4eeb-8d04-a9f10db6150d tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "5d2b56d5-a4e0-4dba-9158-71036d60a40b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.452s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.565298] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315896, 'name': CreateVM_Task} progress is 25%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.595739] env[61868]: DEBUG oslo_vmware.api [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315895, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.837620] env[61868]: DEBUG nova.network.neutron [req-fcc183b8-4341-4d68-a3e5-291770111d12 req-836297f3-ca52-432a-ba94-416b25785b7f service nova] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Updated VIF entry in instance network info cache for port 30ef9d89-a3e3-4d38-8a41-41a35000db0d. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1021.838014] env[61868]: DEBUG nova.network.neutron [req-fcc183b8-4341-4d68-a3e5-291770111d12 req-836297f3-ca52-432a-ba94-416b25785b7f service nova] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Updating instance_info_cache with network_info: [{"id": "30ef9d89-a3e3-4d38-8a41-41a35000db0d", "address": "fa:16:3e:33:d2:c4", "network": {"id": "77f33b52-e4eb-4b72-85d4-5991a2b5c36d", "bridge": "br-int", "label": "tempest-ServersTestJSON-748635995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b520c3ac58074e8d9b0bfafb817244a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30ef9d89-a3", "ovs_interfaceid": "30ef9d89-a3e3-4d38-8a41-41a35000db0d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.905426] env[61868]: DEBUG nova.compute.manager [req-cc1b4c37-058c-451c-9bc0-0b9e987a1b61 req-8da2a20c-b1c6-4fd0-89aa-bec58dbe8b07 service nova] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Received event network-vif-plugged-56a4bffb-5418-4c4c-a6b6-20edaf0e3782 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1021.905761] env[61868]: DEBUG oslo_concurrency.lockutils [req-cc1b4c37-058c-451c-9bc0-0b9e987a1b61 req-8da2a20c-b1c6-4fd0-89aa-bec58dbe8b07 service nova] Acquiring lock "0316bfec-128e-41dc-ad3a-9ba3844b95fa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.906183] env[61868]: DEBUG oslo_concurrency.lockutils [req-cc1b4c37-058c-451c-9bc0-0b9e987a1b61 req-8da2a20c-b1c6-4fd0-89aa-bec58dbe8b07 service nova] Lock "0316bfec-128e-41dc-ad3a-9ba3844b95fa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.906514] env[61868]: DEBUG oslo_concurrency.lockutils [req-cc1b4c37-058c-451c-9bc0-0b9e987a1b61 req-8da2a20c-b1c6-4fd0-89aa-bec58dbe8b07 service nova] Lock "0316bfec-128e-41dc-ad3a-9ba3844b95fa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.906836] env[61868]: DEBUG nova.compute.manager [req-cc1b4c37-058c-451c-9bc0-0b9e987a1b61 req-8da2a20c-b1c6-4fd0-89aa-bec58dbe8b07 service nova] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] No waiting events found dispatching network-vif-plugged-56a4bffb-5418-4c4c-a6b6-20edaf0e3782 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1021.907164] env[61868]: WARNING nova.compute.manager [req-cc1b4c37-058c-451c-9bc0-0b9e987a1b61 req-8da2a20c-b1c6-4fd0-89aa-bec58dbe8b07 service nova] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Received unexpected event network-vif-plugged-56a4bffb-5418-4c4c-a6b6-20edaf0e3782 for instance with vm_state building and task_state spawning. [ 1021.947397] env[61868]: DEBUG nova.network.neutron [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Successfully updated port: 56a4bffb-5418-4c4c-a6b6-20edaf0e3782 {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1022.066059] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315896, 'name': CreateVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.097096] env[61868]: DEBUG oslo_vmware.api [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315895, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.200456] env[61868]: DEBUG nova.network.neutron [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Successfully updated port: 0bbbc97a-2c41-4d5d-a093-fca57a1d543b {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1022.286899] env[61868]: DEBUG nova.compute.manager [req-e9bb6c17-96f8-4bb0-b687-608fe4b7da4c req-aaa519c7-9abc-4381-a1a4-7182be961e6e service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Received event network-vif-plugged-0bbbc97a-2c41-4d5d-a093-fca57a1d543b {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1022.287196] env[61868]: DEBUG oslo_concurrency.lockutils [req-e9bb6c17-96f8-4bb0-b687-608fe4b7da4c req-aaa519c7-9abc-4381-a1a4-7182be961e6e service nova] Acquiring lock "a96cab5a-c053-4e45-96f6-2aba0a819110-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.287426] env[61868]: DEBUG oslo_concurrency.lockutils [req-e9bb6c17-96f8-4bb0-b687-608fe4b7da4c req-aaa519c7-9abc-4381-a1a4-7182be961e6e service nova] Lock "a96cab5a-c053-4e45-96f6-2aba0a819110-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.287606] env[61868]: DEBUG oslo_concurrency.lockutils [req-e9bb6c17-96f8-4bb0-b687-608fe4b7da4c req-aaa519c7-9abc-4381-a1a4-7182be961e6e service nova] Lock "a96cab5a-c053-4e45-96f6-2aba0a819110-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.287782] env[61868]: DEBUG nova.compute.manager [req-e9bb6c17-96f8-4bb0-b687-608fe4b7da4c req-aaa519c7-9abc-4381-a1a4-7182be961e6e service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] No waiting events found dispatching network-vif-plugged-0bbbc97a-2c41-4d5d-a093-fca57a1d543b {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1022.287950] env[61868]: WARNING nova.compute.manager [req-e9bb6c17-96f8-4bb0-b687-608fe4b7da4c req-aaa519c7-9abc-4381-a1a4-7182be961e6e service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Received unexpected event network-vif-plugged-0bbbc97a-2c41-4d5d-a093-fca57a1d543b for instance with vm_state active and task_state None. [ 1022.288526] env[61868]: DEBUG nova.compute.manager [req-e9bb6c17-96f8-4bb0-b687-608fe4b7da4c req-aaa519c7-9abc-4381-a1a4-7182be961e6e service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Received event network-changed-0bbbc97a-2c41-4d5d-a093-fca57a1d543b {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1022.288784] env[61868]: DEBUG nova.compute.manager [req-e9bb6c17-96f8-4bb0-b687-608fe4b7da4c req-aaa519c7-9abc-4381-a1a4-7182be961e6e service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Refreshing instance network info cache due to event network-changed-0bbbc97a-2c41-4d5d-a093-fca57a1d543b. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1022.289056] env[61868]: DEBUG oslo_concurrency.lockutils [req-e9bb6c17-96f8-4bb0-b687-608fe4b7da4c req-aaa519c7-9abc-4381-a1a4-7182be961e6e service nova] Acquiring lock "refresh_cache-a96cab5a-c053-4e45-96f6-2aba0a819110" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.289273] env[61868]: DEBUG oslo_concurrency.lockutils [req-e9bb6c17-96f8-4bb0-b687-608fe4b7da4c req-aaa519c7-9abc-4381-a1a4-7182be961e6e service nova] Acquired lock "refresh_cache-a96cab5a-c053-4e45-96f6-2aba0a819110" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.289495] env[61868]: DEBUG nova.network.neutron [req-e9bb6c17-96f8-4bb0-b687-608fe4b7da4c req-aaa519c7-9abc-4381-a1a4-7182be961e6e service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Refreshing network info cache for port 0bbbc97a-2c41-4d5d-a093-fca57a1d543b {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1022.341045] env[61868]: DEBUG oslo_concurrency.lockutils [req-fcc183b8-4341-4d68-a3e5-291770111d12 req-836297f3-ca52-432a-ba94-416b25785b7f service nova] Releasing lock "refresh_cache-d9e90641-6765-4bda-ab07-3506898f06a9" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.450666] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "refresh_cache-0316bfec-128e-41dc-ad3a-9ba3844b95fa" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.450985] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquired lock "refresh_cache-0316bfec-128e-41dc-ad3a-9ba3844b95fa" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.451162] env[61868]: DEBUG nova.network.neutron [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1022.563203] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315896, 'name': CreateVM_Task, 'duration_secs': 1.102033} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.563357] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1022.564158] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.564522] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.564657] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1022.564906] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1a6a2e9-38fe-4ff2-8e03-7e754bf3ecd9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.577259] env[61868]: DEBUG oslo_vmware.api [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 1022.577259] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52cec2cc-1ca7-aec2-c744-c4ba40f08e88" [ 1022.577259] env[61868]: _type = "Task" [ 1022.577259] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.586876] env[61868]: DEBUG oslo_vmware.api [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52cec2cc-1ca7-aec2-c744-c4ba40f08e88, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.594892] env[61868]: DEBUG oslo_vmware.api [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315895, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.406789} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.595723] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6732cc60-80ef-44f3-89ae-468b81ae48f9/6732cc60-80ef-44f3-89ae-468b81ae48f9.vmdk to [datastore2] c96c4bcb-03f9-4961-aa17-8364aa4cf66e/c96c4bcb-03f9-4961-aa17-8364aa4cf66e.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1022.596540] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64bb8ddf-e4cf-4668-989e-3633d2235c76 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.619594] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] c96c4bcb-03f9-4961-aa17-8364aa4cf66e/c96c4bcb-03f9-4961-aa17-8364aa4cf66e.vmdk or device None with type streamOptimized {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1022.620235] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-590a6ed2-29cf-4b56-aa92-6e9ef3940fb3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.640922] env[61868]: DEBUG oslo_vmware.api [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 1022.640922] env[61868]: value = "task-1315897" [ 1022.640922] env[61868]: _type = "Task" [ 1022.640922] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.649599] env[61868]: DEBUG oslo_vmware.api [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315897, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.703568] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "refresh_cache-a96cab5a-c053-4e45-96f6-2aba0a819110" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.982956] env[61868]: DEBUG nova.network.neutron [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1023.089771] env[61868]: DEBUG oslo_vmware.api [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52cec2cc-1ca7-aec2-c744-c4ba40f08e88, 'name': SearchDatastore_Task, 'duration_secs': 0.010751} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.090111] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.090347] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1023.090613] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1023.090765] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.090950] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1023.091908] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-09e0bc73-39b1-4e3a-8566-e1acc56f0c7f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.099750] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1023.099988] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1023.100729] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5bb0c465-e7ee-4b72-a6af-27324dec11f7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.105895] env[61868]: DEBUG oslo_vmware.api [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 1023.105895] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]528ac55e-0285-e555-4494-405e73599378" [ 1023.105895] env[61868]: _type = "Task" [ 1023.105895] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.115309] env[61868]: DEBUG oslo_vmware.api [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]528ac55e-0285-e555-4494-405e73599378, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.124158] env[61868]: DEBUG nova.network.neutron [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Updating instance_info_cache with network_info: [{"id": "56a4bffb-5418-4c4c-a6b6-20edaf0e3782", "address": "fa:16:3e:58:fc:53", "network": {"id": "c1f5bf03-438f-4acc-93fd-159bed7d4069", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2061742146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a60c3c20950e4156b64c3b4c61b9f0f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24b91e3b-50f1-4a16-b929-942e6b31b2d8", "external-id": "nsx-vlan-transportzone-34", "segmentation_id": 34, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56a4bffb-54", "ovs_interfaceid": "56a4bffb-5418-4c4c-a6b6-20edaf0e3782", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.153148] env[61868]: DEBUG nova.network.neutron [req-e9bb6c17-96f8-4bb0-b687-608fe4b7da4c req-aaa519c7-9abc-4381-a1a4-7182be961e6e service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Added VIF to instance network info cache for port 0bbbc97a-2c41-4d5d-a093-fca57a1d543b. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3546}} [ 1023.153148] env[61868]: DEBUG nova.network.neutron [req-e9bb6c17-96f8-4bb0-b687-608fe4b7da4c req-aaa519c7-9abc-4381-a1a4-7182be961e6e service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Updating instance_info_cache with network_info: [{"id": "cd9efdc6-e6f5-4449-b95d-d1fbec53a295", "address": "fa:16:3e:bf:13:19", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd9efdc6-e6", "ovs_interfaceid": "cd9efdc6-e6f5-4449-b95d-d1fbec53a295", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0bbbc97a-2c41-4d5d-a093-fca57a1d543b", "address": "fa:16:3e:b0:b6:94", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0bbbc97a-2c", "ovs_interfaceid": "0bbbc97a-2c41-4d5d-a093-fca57a1d543b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.158865] env[61868]: DEBUG oslo_vmware.api [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315897, 'name': ReconfigVM_Task, 'duration_secs': 0.329537} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.159238] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Reconfigured VM instance instance-00000060 to attach disk [datastore2] c96c4bcb-03f9-4961-aa17-8364aa4cf66e/c96c4bcb-03f9-4961-aa17-8364aa4cf66e.vmdk or device None with type streamOptimized {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1023.159857] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-30f50992-d515-4c24-a074-76568b4f592d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.167833] env[61868]: DEBUG oslo_vmware.api [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 1023.167833] env[61868]: value = "task-1315898" [ 1023.167833] env[61868]: _type = "Task" [ 1023.167833] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.176362] env[61868]: DEBUG oslo_vmware.api [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315898, 'name': Rename_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.431550] env[61868]: DEBUG oslo_concurrency.lockutils [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "54dacd73-c760-49bd-9baf-ad2012a53264" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.431811] env[61868]: DEBUG oslo_concurrency.lockutils [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "54dacd73-c760-49bd-9baf-ad2012a53264" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.432050] env[61868]: DEBUG nova.compute.manager [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Going to confirm migration 3 {{(pid=61868) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1023.616028] env[61868]: DEBUG oslo_vmware.api [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]528ac55e-0285-e555-4494-405e73599378, 'name': SearchDatastore_Task, 'duration_secs': 0.009083} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.616812] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-92d542e6-4335-4e59-bd0f-4fa478459cab {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.621891] env[61868]: DEBUG oslo_vmware.api [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 1023.621891] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]524b3c17-d15b-71f0-a42e-1db333a68722" [ 1023.621891] env[61868]: _type = "Task" [ 1023.621891] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.626317] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Releasing lock "refresh_cache-0316bfec-128e-41dc-ad3a-9ba3844b95fa" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.626597] env[61868]: DEBUG nova.compute.manager [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Instance network_info: |[{"id": "56a4bffb-5418-4c4c-a6b6-20edaf0e3782", "address": "fa:16:3e:58:fc:53", "network": {"id": "c1f5bf03-438f-4acc-93fd-159bed7d4069", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2061742146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a60c3c20950e4156b64c3b4c61b9f0f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24b91e3b-50f1-4a16-b929-942e6b31b2d8", "external-id": "nsx-vlan-transportzone-34", "segmentation_id": 34, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56a4bffb-54", "ovs_interfaceid": "56a4bffb-5418-4c4c-a6b6-20edaf0e3782", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1023.629829] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:58:fc:53', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24b91e3b-50f1-4a16-b929-942e6b31b2d8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '56a4bffb-5418-4c4c-a6b6-20edaf0e3782', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1023.637451] env[61868]: DEBUG oslo.service.loopingcall [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1023.637740] env[61868]: DEBUG oslo_vmware.api [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]524b3c17-d15b-71f0-a42e-1db333a68722, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.637914] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1023.638135] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-48fe8012-a413-492a-89b1-2722fdbe0d49 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.654373] env[61868]: DEBUG oslo_concurrency.lockutils [req-e9bb6c17-96f8-4bb0-b687-608fe4b7da4c req-aaa519c7-9abc-4381-a1a4-7182be961e6e service nova] Releasing lock "refresh_cache-a96cab5a-c053-4e45-96f6-2aba0a819110" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.654840] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquired lock "refresh_cache-a96cab5a-c053-4e45-96f6-2aba0a819110" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.655033] env[61868]: DEBUG nova.network.neutron [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1023.658168] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1023.658168] env[61868]: value = "task-1315899" [ 1023.658168] env[61868]: _type = "Task" [ 1023.658168] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.668311] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315899, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.676609] env[61868]: DEBUG oslo_vmware.api [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315898, 'name': Rename_Task, 'duration_secs': 0.149536} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.676880] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1023.677392] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ee52edc6-d8a7-4165-85b3-a37d74cfc49a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.683056] env[61868]: DEBUG oslo_vmware.api [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 1023.683056] env[61868]: value = "task-1315900" [ 1023.683056] env[61868]: _type = "Task" [ 1023.683056] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.691434] env[61868]: DEBUG oslo_vmware.api [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315900, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.001942] env[61868]: DEBUG oslo_concurrency.lockutils [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "refresh_cache-54dacd73-c760-49bd-9baf-ad2012a53264" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.002256] env[61868]: DEBUG oslo_concurrency.lockutils [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquired lock "refresh_cache-54dacd73-c760-49bd-9baf-ad2012a53264" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.002478] env[61868]: DEBUG nova.network.neutron [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1024.002730] env[61868]: DEBUG nova.objects.instance [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lazy-loading 'info_cache' on Instance uuid 54dacd73-c760-49bd-9baf-ad2012a53264 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1024.132932] env[61868]: DEBUG oslo_vmware.api [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]524b3c17-d15b-71f0-a42e-1db333a68722, 'name': SearchDatastore_Task, 'duration_secs': 0.009171} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.133221] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.133513] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] d9e90641-6765-4bda-ab07-3506898f06a9/d9e90641-6765-4bda-ab07-3506898f06a9.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1024.133814] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8f3e209e-ae8e-43ae-af2d-755969b010d3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.140661] env[61868]: DEBUG oslo_vmware.api [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 1024.140661] env[61868]: value = "task-1315901" [ 1024.140661] env[61868]: _type = "Task" [ 1024.140661] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.148706] env[61868]: DEBUG oslo_vmware.api [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315901, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.167729] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315899, 'name': CreateVM_Task, 'duration_secs': 0.344248} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.167839] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1024.168512] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.168704] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.169056] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1024.169344] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-283e847f-4677-4e6e-a2ca-dceb6a4d46a9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.174304] env[61868]: DEBUG oslo_vmware.api [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 1024.174304] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5258a63a-56ee-4bba-186c-8d2ebe2d467f" [ 1024.174304] env[61868]: _type = "Task" [ 1024.174304] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.182475] env[61868]: DEBUG oslo_vmware.api [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5258a63a-56ee-4bba-186c-8d2ebe2d467f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.192382] env[61868]: DEBUG oslo_vmware.api [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315900, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.194984] env[61868]: WARNING nova.network.neutron [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] d4f100cd-8de6-4459-a7a6-bc1895b8b14f already exists in list: networks containing: ['d4f100cd-8de6-4459-a7a6-bc1895b8b14f']. ignoring it [ 1024.195264] env[61868]: WARNING nova.network.neutron [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] d4f100cd-8de6-4459-a7a6-bc1895b8b14f already exists in list: networks containing: ['d4f100cd-8de6-4459-a7a6-bc1895b8b14f']. ignoring it [ 1024.195444] env[61868]: WARNING nova.network.neutron [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] 0bbbc97a-2c41-4d5d-a093-fca57a1d543b already exists in list: port_ids containing: ['0bbbc97a-2c41-4d5d-a093-fca57a1d543b']. ignoring it [ 1024.502941] env[61868]: DEBUG nova.network.neutron [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Updating instance_info_cache with network_info: [{"id": "cd9efdc6-e6f5-4449-b95d-d1fbec53a295", "address": "fa:16:3e:bf:13:19", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd9efdc6-e6", "ovs_interfaceid": "cd9efdc6-e6f5-4449-b95d-d1fbec53a295", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0bbbc97a-2c41-4d5d-a093-fca57a1d543b", "address": "fa:16:3e:b0:b6:94", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0bbbc97a-2c", "ovs_interfaceid": "0bbbc97a-2c41-4d5d-a093-fca57a1d543b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.650257] env[61868]: DEBUG oslo_vmware.api [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315901, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.49679} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.650545] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] d9e90641-6765-4bda-ab07-3506898f06a9/d9e90641-6765-4bda-ab07-3506898f06a9.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1024.650756] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1024.651014] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6fb7b05b-c61b-4b60-a45f-f617538a3ed6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.656663] env[61868]: DEBUG oslo_vmware.api [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 1024.656663] env[61868]: value = "task-1315902" [ 1024.656663] env[61868]: _type = "Task" [ 1024.656663] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.663973] env[61868]: DEBUG oslo_vmware.api [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315902, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.683118] env[61868]: DEBUG oslo_vmware.api [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5258a63a-56ee-4bba-186c-8d2ebe2d467f, 'name': SearchDatastore_Task, 'duration_secs': 0.008931} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.683394] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.683663] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1024.684409] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.684409] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.684409] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1024.684568] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6ea9f6c0-2faf-49bf-81f5-8140d71b507c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.693637] env[61868]: DEBUG oslo_vmware.api [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315900, 'name': PowerOnVM_Task, 'duration_secs': 0.566311} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.694609] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1024.694837] env[61868]: INFO nova.compute.manager [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Took 12.91 seconds to spawn the instance on the hypervisor. [ 1024.695052] env[61868]: DEBUG nova.compute.manager [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1024.695328] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1024.695488] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1024.696568] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27e5634e-1f46-4b9e-aa83-6e1e38e6fe00 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.698767] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d7e816b2-bc89-4cd7-9680-6476bb242c2e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.707308] env[61868]: DEBUG oslo_vmware.api [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 1024.707308] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52adac41-a89e-555d-aa9f-d277cb856bd5" [ 1024.707308] env[61868]: _type = "Task" [ 1024.707308] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.714532] env[61868]: DEBUG oslo_vmware.api [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52adac41-a89e-555d-aa9f-d277cb856bd5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.005670] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Releasing lock "refresh_cache-a96cab5a-c053-4e45-96f6-2aba0a819110" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.006427] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "a96cab5a-c053-4e45-96f6-2aba0a819110" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.006602] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquired lock "a96cab5a-c053-4e45-96f6-2aba0a819110" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.009286] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd5a7aed-321a-41cb-a69d-b74dd5187f48 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.027499] env[61868]: DEBUG nova.virt.hardware [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1025.027730] env[61868]: DEBUG nova.virt.hardware [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1025.027889] env[61868]: DEBUG nova.virt.hardware [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1025.028088] env[61868]: DEBUG nova.virt.hardware [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1025.028243] env[61868]: DEBUG nova.virt.hardware [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1025.028393] env[61868]: DEBUG nova.virt.hardware [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1025.028597] env[61868]: DEBUG nova.virt.hardware [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1025.028764] env[61868]: DEBUG nova.virt.hardware [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1025.028927] env[61868]: DEBUG nova.virt.hardware [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1025.029106] env[61868]: DEBUG nova.virt.hardware [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1025.029284] env[61868]: DEBUG nova.virt.hardware [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1025.035528] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Reconfiguring VM to attach interface {{(pid=61868) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1025.035842] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-41e3bb97-6f5c-46f2-b983-6509a927f478 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.054853] env[61868]: DEBUG oslo_vmware.api [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for the task: (returnval){ [ 1025.054853] env[61868]: value = "task-1315903" [ 1025.054853] env[61868]: _type = "Task" [ 1025.054853] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.062435] env[61868]: DEBUG oslo_vmware.api [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315903, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.165725] env[61868]: DEBUG oslo_vmware.api [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315902, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063762} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.165949] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1025.166838] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11bfaded-4dc6-4860-ab27-70dfdf1745b0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.188612] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] d9e90641-6765-4bda-ab07-3506898f06a9/d9e90641-6765-4bda-ab07-3506898f06a9.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1025.191185] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a4047de6-e94d-432b-9f95-d93b680abab3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.217540] env[61868]: DEBUG oslo_vmware.api [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 1025.217540] env[61868]: value = "task-1315904" [ 1025.217540] env[61868]: _type = "Task" [ 1025.217540] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.218022] env[61868]: INFO nova.compute.manager [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Took 19.70 seconds to build instance. [ 1025.226281] env[61868]: DEBUG oslo_vmware.api [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52adac41-a89e-555d-aa9f-d277cb856bd5, 'name': SearchDatastore_Task, 'duration_secs': 0.009348} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.226765] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b91e9f0-f17b-44d7-81e8-801749d371ca {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.231902] env[61868]: DEBUG oslo_vmware.api [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315904, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.235401] env[61868]: DEBUG oslo_vmware.api [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 1025.235401] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]526c857d-dcd3-b3ce-e4d0-73a2a8d3d1d2" [ 1025.235401] env[61868]: _type = "Task" [ 1025.235401] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.243832] env[61868]: DEBUG oslo_vmware.api [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]526c857d-dcd3-b3ce-e4d0-73a2a8d3d1d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.244734] env[61868]: DEBUG nova.network.neutron [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Updating instance_info_cache with network_info: [{"id": "cbd81a07-4122-41bd-bb91-d47b6bc3918d", "address": "fa:16:3e:31:4b:86", "network": {"id": "5b4c5906-b873-4369-855b-e0898ad84812", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-254512637-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "812e115cc1114d0e8536924c4f5556f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcbd81a07-41", "ovs_interfaceid": "cbd81a07-4122-41bd-bb91-d47b6bc3918d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.565117] env[61868]: DEBUG oslo_vmware.api [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315903, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.723715] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6abf46c5-2fa7-42af-8557-47ab2d7ec938 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "c96c4bcb-03f9-4961-aa17-8364aa4cf66e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.215s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.729118] env[61868]: DEBUG oslo_vmware.api [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315904, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.744262] env[61868]: DEBUG oslo_vmware.api [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]526c857d-dcd3-b3ce-e4d0-73a2a8d3d1d2, 'name': SearchDatastore_Task, 'duration_secs': 0.009016} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.744546] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.745045] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 0316bfec-128e-41dc-ad3a-9ba3844b95fa/0316bfec-128e-41dc-ad3a-9ba3844b95fa.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1025.745130] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9dd39431-7400-4612-90dd-b4dee75c4f96 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.747221] env[61868]: DEBUG oslo_concurrency.lockutils [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Releasing lock "refresh_cache-54dacd73-c760-49bd-9baf-ad2012a53264" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.747441] env[61868]: DEBUG nova.objects.instance [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lazy-loading 'migration_context' on Instance uuid 54dacd73-c760-49bd-9baf-ad2012a53264 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1025.753584] env[61868]: DEBUG oslo_vmware.api [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 1025.753584] env[61868]: value = "task-1315905" [ 1025.753584] env[61868]: _type = "Task" [ 1025.753584] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.761798] env[61868]: DEBUG oslo_vmware.api [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315905, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.067280] env[61868]: DEBUG oslo_vmware.api [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315903, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.229316] env[61868]: DEBUG oslo_vmware.api [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315904, 'name': ReconfigVM_Task, 'duration_secs': 0.873802} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.229617] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Reconfigured VM instance instance-00000061 to attach disk [datastore2] d9e90641-6765-4bda-ab07-3506898f06a9/d9e90641-6765-4bda-ab07-3506898f06a9.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1026.230292] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-89fba012-5fa3-4436-89a5-a8211cfff501 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.237211] env[61868]: DEBUG oslo_vmware.api [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 1026.237211] env[61868]: value = "task-1315906" [ 1026.237211] env[61868]: _type = "Task" [ 1026.237211] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.246143] env[61868]: DEBUG oslo_vmware.api [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315906, 'name': Rename_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.249968] env[61868]: DEBUG nova.objects.base [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Object Instance<54dacd73-c760-49bd-9baf-ad2012a53264> lazy-loaded attributes: info_cache,migration_context {{(pid=61868) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1026.250714] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8be2a6ec-6718-4909-abd8-ce43e99c67c7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.271972] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-33f6d62e-8f6e-4e10-84a3-bd70dae9c7db {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.277468] env[61868]: DEBUG oslo_vmware.api [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315905, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.471793} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.278177] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 0316bfec-128e-41dc-ad3a-9ba3844b95fa/0316bfec-128e-41dc-ad3a-9ba3844b95fa.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1026.278447] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1026.278713] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d2c72403-d8f0-4784-880c-2662d0b640b4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.283191] env[61868]: DEBUG oslo_vmware.api [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 1026.283191] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]522f9aca-7965-8932-6654-fb41984df5aa" [ 1026.283191] env[61868]: _type = "Task" [ 1026.283191] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.287876] env[61868]: DEBUG oslo_vmware.api [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 1026.287876] env[61868]: value = "task-1315907" [ 1026.287876] env[61868]: _type = "Task" [ 1026.287876] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.294528] env[61868]: DEBUG oslo_vmware.api [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]522f9aca-7965-8932-6654-fb41984df5aa, 'name': SearchDatastore_Task, 'duration_secs': 0.00702} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.295141] env[61868]: DEBUG oslo_concurrency.lockutils [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.295403] env[61868]: DEBUG oslo_concurrency.lockutils [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.299478] env[61868]: DEBUG oslo_vmware.api [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315907, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.566544] env[61868]: DEBUG oslo_vmware.api [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315903, 'name': ReconfigVM_Task, 'duration_secs': 1.249066} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.567071] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Releasing lock "a96cab5a-c053-4e45-96f6-2aba0a819110" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.567302] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Reconfigured VM to attach interface {{(pid=61868) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1026.748478] env[61868]: DEBUG oslo_vmware.api [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315906, 'name': Rename_Task, 'duration_secs': 0.183111} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.748820] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1026.749065] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6f09ab9d-003c-455b-9e47-8247d16a8ee8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.755998] env[61868]: DEBUG oslo_vmware.api [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 1026.755998] env[61868]: value = "task-1315908" [ 1026.755998] env[61868]: _type = "Task" [ 1026.755998] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.764890] env[61868]: DEBUG oslo_vmware.api [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315908, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.797342] env[61868]: DEBUG oslo_vmware.api [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315907, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.097286} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.799908] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1026.800891] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa2eae37-2992-4265-ac29-990a64f8485b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.830767] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] 0316bfec-128e-41dc-ad3a-9ba3844b95fa/0316bfec-128e-41dc-ad3a-9ba3844b95fa.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1026.834560] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-76dcdbcc-ac47-47f7-8a8a-51648534ef68 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.863475] env[61868]: DEBUG oslo_vmware.api [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 1026.863475] env[61868]: value = "task-1315909" [ 1026.863475] env[61868]: _type = "Task" [ 1026.863475] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.878238] env[61868]: DEBUG oslo_vmware.api [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315909, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.976641] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3a1f6da-e2c7-4d2c-b438-d26790d72393 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.985863] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a60dfb5-71c7-4d05-ad42-44aff623b178 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.716642] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3cf7f0d6-310b-4879-a773-952ed6d3ccad tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "interface-a96cab5a-c053-4e45-96f6-2aba0a819110-0bbbc97a-2c41-4d5d-a093-fca57a1d543b" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 8.643s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.724703] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c9af776-6001-47bf-a6a7-bad976cfa40f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.732625] env[61868]: DEBUG oslo_vmware.api [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315908, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.737267] env[61868]: DEBUG oslo_vmware.api [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315909, 'name': ReconfigVM_Task, 'duration_secs': 0.279327} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.737699] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Reconfigured VM instance instance-00000062 to attach disk [datastore1] 0316bfec-128e-41dc-ad3a-9ba3844b95fa/0316bfec-128e-41dc-ad3a-9ba3844b95fa.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1027.739980] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7417dedf-586f-4cf5-a75b-0f5f8b5a3007 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.743693] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-67879493-56b7-4ace-a692-45becee1813c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.755489] env[61868]: DEBUG nova.compute.provider_tree [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1027.757960] env[61868]: DEBUG oslo_vmware.api [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 1027.757960] env[61868]: value = "task-1315910" [ 1027.757960] env[61868]: _type = "Task" [ 1027.757960] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.765914] env[61868]: DEBUG oslo_vmware.api [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315910, 'name': Rename_Task} progress is 10%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.228601] env[61868]: DEBUG oslo_vmware.api [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315908, 'name': PowerOnVM_Task, 'duration_secs': 1.344715} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.228601] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1028.228782] env[61868]: INFO nova.compute.manager [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Took 9.42 seconds to spawn the instance on the hypervisor. [ 1028.228945] env[61868]: DEBUG nova.compute.manager [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1028.229650] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89d9c759-3bf3-4746-afc8-b4367671f5ee {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.259860] env[61868]: DEBUG nova.scheduler.client.report [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1028.271540] env[61868]: DEBUG oslo_vmware.api [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315910, 'name': Rename_Task, 'duration_secs': 0.13976} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.271797] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1028.272042] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1f6cc92c-058b-4988-a046-41b685690058 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.278439] env[61868]: DEBUG oslo_vmware.api [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 1028.278439] env[61868]: value = "task-1315911" [ 1028.278439] env[61868]: _type = "Task" [ 1028.278439] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.286951] env[61868]: DEBUG oslo_vmware.api [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315911, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.748308] env[61868]: INFO nova.compute.manager [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Took 14.22 seconds to build instance. [ 1028.789840] env[61868]: DEBUG oslo_vmware.api [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315911, 'name': PowerOnVM_Task, 'duration_secs': 0.469765} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.790154] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1028.790354] env[61868]: INFO nova.compute.manager [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Took 7.68 seconds to spawn the instance on the hypervisor. [ 1028.790556] env[61868]: DEBUG nova.compute.manager [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1028.791326] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ebb34ce-bc45-436f-aea6-e5bbf296ec95 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.027217] env[61868]: DEBUG nova.compute.manager [req-be888dd8-e14c-4cce-a569-c4d2035ab402 req-be09a7d8-959d-4ce1-8c86-6a5383c15e0a service nova] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Received event network-changed-56a4bffb-5418-4c4c-a6b6-20edaf0e3782 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1029.027323] env[61868]: DEBUG nova.compute.manager [req-be888dd8-e14c-4cce-a569-c4d2035ab402 req-be09a7d8-959d-4ce1-8c86-6a5383c15e0a service nova] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Refreshing instance network info cache due to event network-changed-56a4bffb-5418-4c4c-a6b6-20edaf0e3782. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1029.027551] env[61868]: DEBUG oslo_concurrency.lockutils [req-be888dd8-e14c-4cce-a569-c4d2035ab402 req-be09a7d8-959d-4ce1-8c86-6a5383c15e0a service nova] Acquiring lock "refresh_cache-0316bfec-128e-41dc-ad3a-9ba3844b95fa" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.027700] env[61868]: DEBUG oslo_concurrency.lockutils [req-be888dd8-e14c-4cce-a569-c4d2035ab402 req-be09a7d8-959d-4ce1-8c86-6a5383c15e0a service nova] Acquired lock "refresh_cache-0316bfec-128e-41dc-ad3a-9ba3844b95fa" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.027864] env[61868]: DEBUG nova.network.neutron [req-be888dd8-e14c-4cce-a569-c4d2035ab402 req-be09a7d8-959d-4ce1-8c86-6a5383c15e0a service nova] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Refreshing network info cache for port 56a4bffb-5418-4c4c-a6b6-20edaf0e3782 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1029.249363] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5e1677da-a611-482c-bfc4-0851e965d773 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "d9e90641-6765-4bda-ab07-3506898f06a9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.730s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.272930] env[61868]: DEBUG oslo_concurrency.lockutils [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.977s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.308880] env[61868]: INFO nova.compute.manager [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Took 13.01 seconds to build instance. [ 1029.710022] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ffb8aa66-ed1f-4e98-aac4-e37111ca41dd tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "c96c4bcb-03f9-4961-aa17-8364aa4cf66e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.710278] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ffb8aa66-ed1f-4e98-aac4-e37111ca41dd tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "c96c4bcb-03f9-4961-aa17-8364aa4cf66e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.710513] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ffb8aa66-ed1f-4e98-aac4-e37111ca41dd tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "c96c4bcb-03f9-4961-aa17-8364aa4cf66e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.710728] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ffb8aa66-ed1f-4e98-aac4-e37111ca41dd tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "c96c4bcb-03f9-4961-aa17-8364aa4cf66e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.710909] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ffb8aa66-ed1f-4e98-aac4-e37111ca41dd tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "c96c4bcb-03f9-4961-aa17-8364aa4cf66e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.713047] env[61868]: INFO nova.compute.manager [None req-ffb8aa66-ed1f-4e98-aac4-e37111ca41dd tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Terminating instance [ 1029.714974] env[61868]: DEBUG nova.compute.manager [None req-ffb8aa66-ed1f-4e98-aac4-e37111ca41dd tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1029.715217] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ffb8aa66-ed1f-4e98-aac4-e37111ca41dd tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1029.716261] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34e9f75a-1c2c-448b-83a4-22599b74b163 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.725609] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffb8aa66-ed1f-4e98-aac4-e37111ca41dd tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1029.725863] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fcd6c1bd-64fd-49e7-a389-119c054c4205 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.729138] env[61868]: DEBUG nova.network.neutron [req-be888dd8-e14c-4cce-a569-c4d2035ab402 req-be09a7d8-959d-4ce1-8c86-6a5383c15e0a service nova] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Updated VIF entry in instance network info cache for port 56a4bffb-5418-4c4c-a6b6-20edaf0e3782. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1029.729465] env[61868]: DEBUG nova.network.neutron [req-be888dd8-e14c-4cce-a569-c4d2035ab402 req-be09a7d8-959d-4ce1-8c86-6a5383c15e0a service nova] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Updating instance_info_cache with network_info: [{"id": "56a4bffb-5418-4c4c-a6b6-20edaf0e3782", "address": "fa:16:3e:58:fc:53", "network": {"id": "c1f5bf03-438f-4acc-93fd-159bed7d4069", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2061742146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a60c3c20950e4156b64c3b4c61b9f0f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24b91e3b-50f1-4a16-b929-942e6b31b2d8", "external-id": "nsx-vlan-transportzone-34", "segmentation_id": 34, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56a4bffb-54", "ovs_interfaceid": "56a4bffb-5418-4c4c-a6b6-20edaf0e3782", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.731621] env[61868]: DEBUG oslo_vmware.api [None req-ffb8aa66-ed1f-4e98-aac4-e37111ca41dd tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 1029.731621] env[61868]: value = "task-1315912" [ 1029.731621] env[61868]: _type = "Task" [ 1029.731621] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.740317] env[61868]: DEBUG oslo_vmware.api [None req-ffb8aa66-ed1f-4e98-aac4-e37111ca41dd tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315912, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.760732] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e85de296-a39d-423a-a0c2-369ae15a0786 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "interface-a96cab5a-c053-4e45-96f6-2aba0a819110-0bbbc97a-2c41-4d5d-a093-fca57a1d543b" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.761013] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e85de296-a39d-423a-a0c2-369ae15a0786 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "interface-a96cab5a-c053-4e45-96f6-2aba0a819110-0bbbc97a-2c41-4d5d-a093-fca57a1d543b" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.811028] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0f396b1a-e502-4821-bda9-672661c6ea79 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "0316bfec-128e-41dc-ad3a-9ba3844b95fa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.518s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.831779] env[61868]: INFO nova.scheduler.client.report [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Deleted allocation for migration 8b2b7aa9-6eb5-4898-bdc4-cd3c2b1acdd8 [ 1030.056222] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8a8748ac-a070-45b9-b2c8-e2562a82a4d0 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "eeeec752-7e99-471d-9425-c4ccf4f6d0a5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.056502] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8a8748ac-a070-45b9-b2c8-e2562a82a4d0 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "eeeec752-7e99-471d-9425-c4ccf4f6d0a5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.057031] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8a8748ac-a070-45b9-b2c8-e2562a82a4d0 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "eeeec752-7e99-471d-9425-c4ccf4f6d0a5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.057031] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8a8748ac-a070-45b9-b2c8-e2562a82a4d0 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "eeeec752-7e99-471d-9425-c4ccf4f6d0a5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.057138] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8a8748ac-a070-45b9-b2c8-e2562a82a4d0 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "eeeec752-7e99-471d-9425-c4ccf4f6d0a5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.059945] env[61868]: INFO nova.compute.manager [None req-8a8748ac-a070-45b9-b2c8-e2562a82a4d0 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Terminating instance [ 1030.061208] env[61868]: DEBUG nova.compute.manager [None req-8a8748ac-a070-45b9-b2c8-e2562a82a4d0 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1030.061560] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-8a8748ac-a070-45b9-b2c8-e2562a82a4d0 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1030.062428] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1c6d98a-f9ec-4e92-85e1-64e90c824ae0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.070489] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a8748ac-a070-45b9-b2c8-e2562a82a4d0 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1030.070717] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1343ea2e-cb39-42a5-b088-ad6ff8c91ae6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.076894] env[61868]: DEBUG oslo_vmware.api [None req-8a8748ac-a070-45b9-b2c8-e2562a82a4d0 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 1030.076894] env[61868]: value = "task-1315913" [ 1030.076894] env[61868]: _type = "Task" [ 1030.076894] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.084324] env[61868]: DEBUG oslo_vmware.api [None req-8a8748ac-a070-45b9-b2c8-e2562a82a4d0 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315913, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.232988] env[61868]: DEBUG oslo_concurrency.lockutils [req-be888dd8-e14c-4cce-a569-c4d2035ab402 req-be09a7d8-959d-4ce1-8c86-6a5383c15e0a service nova] Releasing lock "refresh_cache-0316bfec-128e-41dc-ad3a-9ba3844b95fa" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.241609] env[61868]: DEBUG oslo_vmware.api [None req-ffb8aa66-ed1f-4e98-aac4-e37111ca41dd tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315912, 'name': PowerOffVM_Task, 'duration_secs': 0.256853} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.241884] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffb8aa66-ed1f-4e98-aac4-e37111ca41dd tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1030.242068] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ffb8aa66-ed1f-4e98-aac4-e37111ca41dd tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1030.242316] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d6a64efa-e9d5-47cd-9133-92e273499eb0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.263645] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e85de296-a39d-423a-a0c2-369ae15a0786 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "a96cab5a-c053-4e45-96f6-2aba0a819110" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.263830] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e85de296-a39d-423a-a0c2-369ae15a0786 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquired lock "a96cab5a-c053-4e45-96f6-2aba0a819110" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.264606] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8afea553-643e-4c17-960c-bd699a23f0e2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.281316] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53a45326-c763-4b53-a8a8-8be1dc6725e9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.306345] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-e85de296-a39d-423a-a0c2-369ae15a0786 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Reconfiguring VM to detach interface {{(pid=61868) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1030.307886] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eda417b9-d960-4b01-bd7b-829d1de897c2 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "d9e90641-6765-4bda-ab07-3506898f06a9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.308137] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eda417b9-d960-4b01-bd7b-829d1de897c2 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "d9e90641-6765-4bda-ab07-3506898f06a9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.308340] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eda417b9-d960-4b01-bd7b-829d1de897c2 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "d9e90641-6765-4bda-ab07-3506898f06a9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.308529] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eda417b9-d960-4b01-bd7b-829d1de897c2 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "d9e90641-6765-4bda-ab07-3506898f06a9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.308699] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eda417b9-d960-4b01-bd7b-829d1de897c2 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "d9e90641-6765-4bda-ab07-3506898f06a9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.310265] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-07a8c840-9471-4e38-87a7-8e18cb9a77dc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.324418] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ffb8aa66-ed1f-4e98-aac4-e37111ca41dd tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1030.324643] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ffb8aa66-ed1f-4e98-aac4-e37111ca41dd tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Deleting contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1030.324845] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-ffb8aa66-ed1f-4e98-aac4-e37111ca41dd tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Deleting the datastore file [datastore2] c96c4bcb-03f9-4961-aa17-8364aa4cf66e {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1030.325399] env[61868]: INFO nova.compute.manager [None req-eda417b9-d960-4b01-bd7b-829d1de897c2 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Terminating instance [ 1030.327084] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-07aaae0f-1921-4ece-b5a0-56d86e4ede46 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.329042] env[61868]: DEBUG nova.compute.manager [None req-eda417b9-d960-4b01-bd7b-829d1de897c2 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1030.329241] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-eda417b9-d960-4b01-bd7b-829d1de897c2 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1030.330319] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a04e9137-b539-4085-8454-a67cd77ef423 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.335104] env[61868]: DEBUG oslo_vmware.api [None req-e85de296-a39d-423a-a0c2-369ae15a0786 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for the task: (returnval){ [ 1030.335104] env[61868]: value = "task-1315915" [ 1030.335104] env[61868]: _type = "Task" [ 1030.335104] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.340415] env[61868]: DEBUG oslo_concurrency.lockutils [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "54dacd73-c760-49bd-9baf-ad2012a53264" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.909s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.341340] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-eda417b9-d960-4b01-bd7b-829d1de897c2 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1030.342371] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-797a3bc0-d3ca-49c7-a3a0-63b854532718 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.343869] env[61868]: DEBUG oslo_vmware.api [None req-ffb8aa66-ed1f-4e98-aac4-e37111ca41dd tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 1030.343869] env[61868]: value = "task-1315916" [ 1030.343869] env[61868]: _type = "Task" [ 1030.343869] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.347350] env[61868]: DEBUG oslo_vmware.api [None req-e85de296-a39d-423a-a0c2-369ae15a0786 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315915, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.351498] env[61868]: DEBUG oslo_vmware.api [None req-eda417b9-d960-4b01-bd7b-829d1de897c2 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 1030.351498] env[61868]: value = "task-1315917" [ 1030.351498] env[61868]: _type = "Task" [ 1030.351498] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.360273] env[61868]: DEBUG oslo_vmware.api [None req-ffb8aa66-ed1f-4e98-aac4-e37111ca41dd tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315916, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.365904] env[61868]: DEBUG oslo_vmware.api [None req-eda417b9-d960-4b01-bd7b-829d1de897c2 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315917, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.589752] env[61868]: DEBUG oslo_vmware.api [None req-8a8748ac-a070-45b9-b2c8-e2562a82a4d0 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315913, 'name': PowerOffVM_Task, 'duration_secs': 0.166609} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.589752] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a8748ac-a070-45b9-b2c8-e2562a82a4d0 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1030.589930] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-8a8748ac-a070-45b9-b2c8-e2562a82a4d0 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1030.590198] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fbf16eb4-8425-43c4-8370-ae3ec98f461b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.647996] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-8a8748ac-a070-45b9-b2c8-e2562a82a4d0 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1030.648293] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-8a8748ac-a070-45b9-b2c8-e2562a82a4d0 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1030.648482] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a8748ac-a070-45b9-b2c8-e2562a82a4d0 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Deleting the datastore file [datastore1] eeeec752-7e99-471d-9425-c4ccf4f6d0a5 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1030.648753] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eeeeb6ab-6f7a-4f7f-94c8-3b046d7dffa0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.655542] env[61868]: DEBUG oslo_vmware.api [None req-8a8748ac-a070-45b9-b2c8-e2562a82a4d0 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 1030.655542] env[61868]: value = "task-1315919" [ 1030.655542] env[61868]: _type = "Task" [ 1030.655542] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.663181] env[61868]: DEBUG oslo_vmware.api [None req-8a8748ac-a070-45b9-b2c8-e2562a82a4d0 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315919, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.844374] env[61868]: DEBUG oslo_vmware.api [None req-e85de296-a39d-423a-a0c2-369ae15a0786 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315915, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.858352] env[61868]: DEBUG oslo_vmware.api [None req-ffb8aa66-ed1f-4e98-aac4-e37111ca41dd tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315916, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140659} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.858917] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-ffb8aa66-ed1f-4e98-aac4-e37111ca41dd tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1030.859171] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ffb8aa66-ed1f-4e98-aac4-e37111ca41dd tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Deleted contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1030.859495] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ffb8aa66-ed1f-4e98-aac4-e37111ca41dd tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1030.859679] env[61868]: INFO nova.compute.manager [None req-ffb8aa66-ed1f-4e98-aac4-e37111ca41dd tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1030.859913] env[61868]: DEBUG oslo.service.loopingcall [None req-ffb8aa66-ed1f-4e98-aac4-e37111ca41dd tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1030.860132] env[61868]: DEBUG nova.compute.manager [-] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1030.860226] env[61868]: DEBUG nova.network.neutron [-] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1030.864334] env[61868]: DEBUG oslo_vmware.api [None req-eda417b9-d960-4b01-bd7b-829d1de897c2 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315917, 'name': PowerOffVM_Task, 'duration_secs': 0.21607} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.864998] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-eda417b9-d960-4b01-bd7b-829d1de897c2 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1030.865266] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-eda417b9-d960-4b01-bd7b-829d1de897c2 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1030.865509] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-356feee4-1ed5-489d-bacb-e6e07617bf9b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.930880] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-eda417b9-d960-4b01-bd7b-829d1de897c2 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1030.931174] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-eda417b9-d960-4b01-bd7b-829d1de897c2 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Deleting contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1030.931342] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-eda417b9-d960-4b01-bd7b-829d1de897c2 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Deleting the datastore file [datastore2] d9e90641-6765-4bda-ab07-3506898f06a9 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1030.931616] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d4a24b29-166c-44dd-b802-06a308a26a21 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.937829] env[61868]: DEBUG oslo_vmware.api [None req-eda417b9-d960-4b01-bd7b-829d1de897c2 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 1030.937829] env[61868]: value = "task-1315921" [ 1030.937829] env[61868]: _type = "Task" [ 1030.937829] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.950017] env[61868]: DEBUG oslo_vmware.api [None req-eda417b9-d960-4b01-bd7b-829d1de897c2 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315921, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.141823] env[61868]: DEBUG nova.compute.manager [req-d0e6ac8d-f8ad-48ce-9011-9a9792e345f8 req-e1f81c98-08d6-45b3-bafb-85a07338fe14 service nova] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Received event network-vif-deleted-b0fdd6cb-31bd-4a5d-95cd-64d82af61df9 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1031.142016] env[61868]: INFO nova.compute.manager [req-d0e6ac8d-f8ad-48ce-9011-9a9792e345f8 req-e1f81c98-08d6-45b3-bafb-85a07338fe14 service nova] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Neutron deleted interface b0fdd6cb-31bd-4a5d-95cd-64d82af61df9; detaching it from the instance and deleting it from the info cache [ 1031.142237] env[61868]: DEBUG nova.network.neutron [req-d0e6ac8d-f8ad-48ce-9011-9a9792e345f8 req-e1f81c98-08d6-45b3-bafb-85a07338fe14 service nova] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.169234] env[61868]: DEBUG oslo_vmware.api [None req-8a8748ac-a070-45b9-b2c8-e2562a82a4d0 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315919, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.369477} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.169511] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a8748ac-a070-45b9-b2c8-e2562a82a4d0 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1031.169756] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-8a8748ac-a070-45b9-b2c8-e2562a82a4d0 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1031.169945] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-8a8748ac-a070-45b9-b2c8-e2562a82a4d0 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1031.170139] env[61868]: INFO nova.compute.manager [None req-8a8748ac-a070-45b9-b2c8-e2562a82a4d0 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1031.170381] env[61868]: DEBUG oslo.service.loopingcall [None req-8a8748ac-a070-45b9-b2c8-e2562a82a4d0 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1031.170857] env[61868]: DEBUG nova.compute.manager [-] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1031.170954] env[61868]: DEBUG nova.network.neutron [-] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1031.243326] env[61868]: DEBUG oslo_concurrency.lockutils [None req-84e02b16-e99c-417d-8084-56b7f32bccb3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "0316bfec-128e-41dc-ad3a-9ba3844b95fa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.243759] env[61868]: DEBUG oslo_concurrency.lockutils [None req-84e02b16-e99c-417d-8084-56b7f32bccb3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "0316bfec-128e-41dc-ad3a-9ba3844b95fa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.244133] env[61868]: DEBUG oslo_concurrency.lockutils [None req-84e02b16-e99c-417d-8084-56b7f32bccb3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "0316bfec-128e-41dc-ad3a-9ba3844b95fa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.244461] env[61868]: DEBUG oslo_concurrency.lockutils [None req-84e02b16-e99c-417d-8084-56b7f32bccb3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "0316bfec-128e-41dc-ad3a-9ba3844b95fa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.244762] env[61868]: DEBUG oslo_concurrency.lockutils [None req-84e02b16-e99c-417d-8084-56b7f32bccb3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "0316bfec-128e-41dc-ad3a-9ba3844b95fa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.248155] env[61868]: INFO nova.compute.manager [None req-84e02b16-e99c-417d-8084-56b7f32bccb3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Terminating instance [ 1031.250648] env[61868]: DEBUG nova.compute.manager [None req-84e02b16-e99c-417d-8084-56b7f32bccb3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1031.250964] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-84e02b16-e99c-417d-8084-56b7f32bccb3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1031.252258] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76a26df0-4407-496e-b067-6bef951ccc18 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.262499] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-84e02b16-e99c-417d-8084-56b7f32bccb3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1031.262867] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8793e34e-2310-44e1-b723-f934b9529e7f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.270607] env[61868]: DEBUG oslo_vmware.api [None req-84e02b16-e99c-417d-8084-56b7f32bccb3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 1031.270607] env[61868]: value = "task-1315922" [ 1031.270607] env[61868]: _type = "Task" [ 1031.270607] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.283804] env[61868]: DEBUG oslo_vmware.api [None req-84e02b16-e99c-417d-8084-56b7f32bccb3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315922, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.345592] env[61868]: DEBUG oslo_vmware.api [None req-e85de296-a39d-423a-a0c2-369ae15a0786 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315915, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.448590] env[61868]: DEBUG oslo_vmware.api [None req-eda417b9-d960-4b01-bd7b-829d1de897c2 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315921, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.354208} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.448590] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-eda417b9-d960-4b01-bd7b-829d1de897c2 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1031.448590] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-eda417b9-d960-4b01-bd7b-829d1de897c2 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Deleted contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1031.448817] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-eda417b9-d960-4b01-bd7b-829d1de897c2 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1031.448910] env[61868]: INFO nova.compute.manager [None req-eda417b9-d960-4b01-bd7b-829d1de897c2 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1031.449168] env[61868]: DEBUG oslo.service.loopingcall [None req-eda417b9-d960-4b01-bd7b-829d1de897c2 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1031.449381] env[61868]: DEBUG nova.compute.manager [-] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1031.449476] env[61868]: DEBUG nova.network.neutron [-] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1031.451345] env[61868]: DEBUG oslo_concurrency.lockutils [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "54dacd73-c760-49bd-9baf-ad2012a53264" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.451712] env[61868]: DEBUG oslo_concurrency.lockutils [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "54dacd73-c760-49bd-9baf-ad2012a53264" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.451810] env[61868]: DEBUG oslo_concurrency.lockutils [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "54dacd73-c760-49bd-9baf-ad2012a53264-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.452010] env[61868]: DEBUG oslo_concurrency.lockutils [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "54dacd73-c760-49bd-9baf-ad2012a53264-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.452191] env[61868]: DEBUG oslo_concurrency.lockutils [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "54dacd73-c760-49bd-9baf-ad2012a53264-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.456155] env[61868]: INFO nova.compute.manager [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Terminating instance [ 1031.458055] env[61868]: DEBUG nova.compute.manager [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1031.458260] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1031.459479] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08768993-ee0c-4dae-9a03-2787a117ddd7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.469024] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1031.469263] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7e897d6c-31bf-4365-bf1e-cfb06ac4cf2b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.475547] env[61868]: DEBUG oslo_vmware.api [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 1031.475547] env[61868]: value = "task-1315923" [ 1031.475547] env[61868]: _type = "Task" [ 1031.475547] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.484040] env[61868]: DEBUG oslo_vmware.api [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315923, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.620746] env[61868]: DEBUG nova.network.neutron [-] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.647428] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d950e702-bc60-4b4e-891c-3e549a3429d1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.657919] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7eb10cc-4cf3-40d0-bc4f-e92f4c6e6563 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.685364] env[61868]: DEBUG nova.compute.manager [req-d0e6ac8d-f8ad-48ce-9011-9a9792e345f8 req-e1f81c98-08d6-45b3-bafb-85a07338fe14 service nova] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Detach interface failed, port_id=b0fdd6cb-31bd-4a5d-95cd-64d82af61df9, reason: Instance c96c4bcb-03f9-4961-aa17-8364aa4cf66e could not be found. {{(pid=61868) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1031.710472] env[61868]: DEBUG nova.compute.manager [req-2d38521e-4778-4433-a402-bee415217e0c req-6b973f6d-bf76-406e-aa20-7adedcb4d9b6 service nova] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Received event network-vif-deleted-30ef9d89-a3e3-4d38-8a41-41a35000db0d {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1031.710814] env[61868]: INFO nova.compute.manager [req-2d38521e-4778-4433-a402-bee415217e0c req-6b973f6d-bf76-406e-aa20-7adedcb4d9b6 service nova] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Neutron deleted interface 30ef9d89-a3e3-4d38-8a41-41a35000db0d; detaching it from the instance and deleting it from the info cache [ 1031.711066] env[61868]: DEBUG nova.network.neutron [req-2d38521e-4778-4433-a402-bee415217e0c req-6b973f6d-bf76-406e-aa20-7adedcb4d9b6 service nova] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.781501] env[61868]: DEBUG oslo_vmware.api [None req-84e02b16-e99c-417d-8084-56b7f32bccb3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315922, 'name': PowerOffVM_Task, 'duration_secs': 0.192755} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.781774] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-84e02b16-e99c-417d-8084-56b7f32bccb3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1031.781945] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-84e02b16-e99c-417d-8084-56b7f32bccb3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1031.782206] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2ae92a05-75fb-49bb-b238-4443f52269be {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.846311] env[61868]: DEBUG oslo_vmware.api [None req-e85de296-a39d-423a-a0c2-369ae15a0786 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315915, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.847354] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-84e02b16-e99c-417d-8084-56b7f32bccb3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1031.847542] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-84e02b16-e99c-417d-8084-56b7f32bccb3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1031.847722] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-84e02b16-e99c-417d-8084-56b7f32bccb3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Deleting the datastore file [datastore1] 0316bfec-128e-41dc-ad3a-9ba3844b95fa {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1031.847964] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2e0106f9-dccf-4382-9b1d-2acf9718f421 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.853920] env[61868]: DEBUG oslo_vmware.api [None req-84e02b16-e99c-417d-8084-56b7f32bccb3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for the task: (returnval){ [ 1031.853920] env[61868]: value = "task-1315925" [ 1031.853920] env[61868]: _type = "Task" [ 1031.853920] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.862342] env[61868]: DEBUG oslo_vmware.api [None req-84e02b16-e99c-417d-8084-56b7f32bccb3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315925, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.985722] env[61868]: DEBUG oslo_vmware.api [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315923, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.105441] env[61868]: DEBUG nova.network.neutron [-] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.123120] env[61868]: INFO nova.compute.manager [-] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Took 1.26 seconds to deallocate network for instance. [ 1032.188416] env[61868]: DEBUG nova.network.neutron [-] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.213923] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4660e9d0-a0f6-48e6-b2eb-054c20ff0612 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.223347] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57b23781-4cac-4756-8c29-bd45fe556df1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.248628] env[61868]: DEBUG nova.compute.manager [req-2d38521e-4778-4433-a402-bee415217e0c req-6b973f6d-bf76-406e-aa20-7adedcb4d9b6 service nova] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Detach interface failed, port_id=30ef9d89-a3e3-4d38-8a41-41a35000db0d, reason: Instance d9e90641-6765-4bda-ab07-3506898f06a9 could not be found. {{(pid=61868) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1032.345980] env[61868]: DEBUG oslo_vmware.api [None req-e85de296-a39d-423a-a0c2-369ae15a0786 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315915, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.361834] env[61868]: DEBUG oslo_vmware.api [None req-84e02b16-e99c-417d-8084-56b7f32bccb3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Task: {'id': task-1315925, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140474} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.362090] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-84e02b16-e99c-417d-8084-56b7f32bccb3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1032.362278] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-84e02b16-e99c-417d-8084-56b7f32bccb3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1032.362459] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-84e02b16-e99c-417d-8084-56b7f32bccb3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1032.362636] env[61868]: INFO nova.compute.manager [None req-84e02b16-e99c-417d-8084-56b7f32bccb3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1032.362866] env[61868]: DEBUG oslo.service.loopingcall [None req-84e02b16-e99c-417d-8084-56b7f32bccb3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1032.363074] env[61868]: DEBUG nova.compute.manager [-] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1032.363176] env[61868]: DEBUG nova.network.neutron [-] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1032.485561] env[61868]: DEBUG oslo_vmware.api [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315923, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.607731] env[61868]: INFO nova.compute.manager [-] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Took 1.44 seconds to deallocate network for instance. [ 1032.630437] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ffb8aa66-ed1f-4e98-aac4-e37111ca41dd tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.630712] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ffb8aa66-ed1f-4e98-aac4-e37111ca41dd tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.630938] env[61868]: DEBUG nova.objects.instance [None req-ffb8aa66-ed1f-4e98-aac4-e37111ca41dd tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lazy-loading 'resources' on Instance uuid c96c4bcb-03f9-4961-aa17-8364aa4cf66e {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1032.694202] env[61868]: INFO nova.compute.manager [-] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Took 1.24 seconds to deallocate network for instance. [ 1032.847059] env[61868]: DEBUG oslo_vmware.api [None req-e85de296-a39d-423a-a0c2-369ae15a0786 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315915, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.986920] env[61868]: DEBUG oslo_vmware.api [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315923, 'name': PowerOffVM_Task, 'duration_secs': 1.24339} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.989377] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1032.989377] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1032.989377] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fe0a2834-4611-44f2-a0c5-6bbbb641cf7d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.058758] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1033.058758] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Deleting contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1033.058758] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Deleting the datastore file [datastore2] 54dacd73-c760-49bd-9baf-ad2012a53264 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1033.058953] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b6c1c350-9c6c-4ab1-96a4-2cc3ef2831b3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.065725] env[61868]: DEBUG oslo_vmware.api [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for the task: (returnval){ [ 1033.065725] env[61868]: value = "task-1315927" [ 1033.065725] env[61868]: _type = "Task" [ 1033.065725] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.073956] env[61868]: DEBUG oslo_vmware.api [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315927, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.096451] env[61868]: DEBUG nova.network.neutron [-] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.113782] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8a8748ac-a070-45b9-b2c8-e2562a82a4d0 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.165645] env[61868]: DEBUG nova.compute.manager [req-33dc1009-a34e-4afd-83c4-9c9a62f3a9c1 req-9df797aa-2a93-41d9-a079-c38355f2077a service nova] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Received event network-vif-deleted-8f5f51e1-8b43-4ef6-b9b8-19c6b4e89e6e {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1033.200952] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eda417b9-d960-4b01-bd7b-829d1de897c2 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.250788] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-679fe066-56e3-46d5-ac78-dc07565bd1e5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.257295] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbd2cd37-edb1-4fe4-8045-9d4387fab86c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.287048] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fef4ef44-add1-4cc4-8343-e42c7ddff37c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.294159] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bccd4e1-c5df-4283-bbe6-96553b0ba0c0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.308040] env[61868]: DEBUG nova.compute.provider_tree [None req-ffb8aa66-ed1f-4e98-aac4-e37111ca41dd tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1033.349184] env[61868]: DEBUG oslo_vmware.api [None req-e85de296-a39d-423a-a0c2-369ae15a0786 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315915, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.575905] env[61868]: DEBUG oslo_vmware.api [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Task: {'id': task-1315927, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.156017} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.576179] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1033.576371] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Deleted contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1033.576550] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1033.576722] env[61868]: INFO nova.compute.manager [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Took 2.12 seconds to destroy the instance on the hypervisor. [ 1033.577138] env[61868]: DEBUG oslo.service.loopingcall [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1033.577348] env[61868]: DEBUG nova.compute.manager [-] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1033.577441] env[61868]: DEBUG nova.network.neutron [-] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1033.599642] env[61868]: INFO nova.compute.manager [-] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Took 1.24 seconds to deallocate network for instance. [ 1033.737019] env[61868]: DEBUG nova.compute.manager [req-cba041a2-a757-444f-bfac-465342a9f3a6 req-b17acb31-7c38-43b9-86e2-949522ff80a2 service nova] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Received event network-vif-deleted-56a4bffb-5418-4c4c-a6b6-20edaf0e3782 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1033.811496] env[61868]: DEBUG nova.scheduler.client.report [None req-ffb8aa66-ed1f-4e98-aac4-e37111ca41dd tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1033.847885] env[61868]: DEBUG oslo_vmware.api [None req-e85de296-a39d-423a-a0c2-369ae15a0786 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315915, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.106739] env[61868]: DEBUG oslo_concurrency.lockutils [None req-84e02b16-e99c-417d-8084-56b7f32bccb3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.295070] env[61868]: DEBUG nova.network.neutron [-] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.315979] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ffb8aa66-ed1f-4e98-aac4-e37111ca41dd tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.685s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.318414] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8a8748ac-a070-45b9-b2c8-e2562a82a4d0 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.205s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.318689] env[61868]: DEBUG nova.objects.instance [None req-8a8748ac-a070-45b9-b2c8-e2562a82a4d0 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lazy-loading 'resources' on Instance uuid eeeec752-7e99-471d-9425-c4ccf4f6d0a5 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1034.340104] env[61868]: INFO nova.scheduler.client.report [None req-ffb8aa66-ed1f-4e98-aac4-e37111ca41dd tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Deleted allocations for instance c96c4bcb-03f9-4961-aa17-8364aa4cf66e [ 1034.354125] env[61868]: DEBUG oslo_vmware.api [None req-e85de296-a39d-423a-a0c2-369ae15a0786 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315915, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.797094] env[61868]: INFO nova.compute.manager [-] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Took 1.22 seconds to deallocate network for instance. [ 1034.853266] env[61868]: DEBUG oslo_vmware.api [None req-e85de296-a39d-423a-a0c2-369ae15a0786 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315915, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.856006] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ffb8aa66-ed1f-4e98-aac4-e37111ca41dd tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "c96c4bcb-03f9-4961-aa17-8364aa4cf66e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.146s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.932120] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68be4107-a5cb-42a1-b3d7-0a9bbcdf980e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.939323] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e969fbcb-4d25-4b55-8cf3-a3ad07c8c481 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.971137] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63f17b5d-920e-4d4f-9391-efd7bbbd95ea {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.978909] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15a643a1-d750-4832-b40a-10f0aaad99bf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.991960] env[61868]: DEBUG nova.compute.provider_tree [None req-8a8748ac-a070-45b9-b2c8-e2562a82a4d0 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1035.304582] env[61868]: DEBUG oslo_concurrency.lockutils [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.352547] env[61868]: DEBUG oslo_vmware.api [None req-e85de296-a39d-423a-a0c2-369ae15a0786 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315915, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.495500] env[61868]: DEBUG nova.scheduler.client.report [None req-8a8748ac-a070-45b9-b2c8-e2562a82a4d0 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1035.519795] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8b0db86f-df1c-40ac-889f-6d9f939a48bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "cbeeb7b1-166b-49f7-88c5-6ed05c9dc759" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.520066] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8b0db86f-df1c-40ac-889f-6d9f939a48bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "cbeeb7b1-166b-49f7-88c5-6ed05c9dc759" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.520344] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8b0db86f-df1c-40ac-889f-6d9f939a48bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "cbeeb7b1-166b-49f7-88c5-6ed05c9dc759-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.520462] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8b0db86f-df1c-40ac-889f-6d9f939a48bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "cbeeb7b1-166b-49f7-88c5-6ed05c9dc759-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.520850] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8b0db86f-df1c-40ac-889f-6d9f939a48bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "cbeeb7b1-166b-49f7-88c5-6ed05c9dc759-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.522973] env[61868]: INFO nova.compute.manager [None req-8b0db86f-df1c-40ac-889f-6d9f939a48bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Terminating instance [ 1035.524713] env[61868]: DEBUG nova.compute.manager [None req-8b0db86f-df1c-40ac-889f-6d9f939a48bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1035.524877] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-8b0db86f-df1c-40ac-889f-6d9f939a48bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1035.525752] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-311eeeec-8e5a-4af2-9712-26f321656c42 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.538471] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b0db86f-df1c-40ac-889f-6d9f939a48bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1035.538712] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bc28d266-98a7-415d-bb15-beb1b44fe8b9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.544268] env[61868]: DEBUG oslo_vmware.api [None req-8b0db86f-df1c-40ac-889f-6d9f939a48bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 1035.544268] env[61868]: value = "task-1315928" [ 1035.544268] env[61868]: _type = "Task" [ 1035.544268] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.552141] env[61868]: DEBUG oslo_vmware.api [None req-8b0db86f-df1c-40ac-889f-6d9f939a48bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315928, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.761597] env[61868]: DEBUG nova.compute.manager [req-30de24f3-e00b-462d-886a-f108ada9b1ff req-65b1f0a3-9ab5-41d4-b4c5-f227b288dbe9 service nova] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Received event network-vif-deleted-cbd81a07-4122-41bd-bb91-d47b6bc3918d {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1035.852308] env[61868]: DEBUG oslo_vmware.api [None req-e85de296-a39d-423a-a0c2-369ae15a0786 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315915, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.000102] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8a8748ac-a070-45b9-b2c8-e2562a82a4d0 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.682s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.003740] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eda417b9-d960-4b01-bd7b-829d1de897c2 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.803s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.003997] env[61868]: DEBUG nova.objects.instance [None req-eda417b9-d960-4b01-bd7b-829d1de897c2 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lazy-loading 'resources' on Instance uuid d9e90641-6765-4bda-ab07-3506898f06a9 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1036.008322] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.008453] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.018348] env[61868]: INFO nova.scheduler.client.report [None req-8a8748ac-a070-45b9-b2c8-e2562a82a4d0 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Deleted allocations for instance eeeec752-7e99-471d-9425-c4ccf4f6d0a5 [ 1036.053843] env[61868]: DEBUG oslo_vmware.api [None req-8b0db86f-df1c-40ac-889f-6d9f939a48bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315928, 'name': PowerOffVM_Task, 'duration_secs': 0.187704} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.054168] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b0db86f-df1c-40ac-889f-6d9f939a48bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1036.054351] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-8b0db86f-df1c-40ac-889f-6d9f939a48bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1036.054604] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-aa389876-a184-42d2-9ca3-36b5b1764d07 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.122779] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-8b0db86f-df1c-40ac-889f-6d9f939a48bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1036.123160] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-8b0db86f-df1c-40ac-889f-6d9f939a48bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1036.123373] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b0db86f-df1c-40ac-889f-6d9f939a48bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Deleting the datastore file [datastore1] cbeeb7b1-166b-49f7-88c5-6ed05c9dc759 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1036.123645] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c2db2e8d-f4a2-48ad-b265-18641a88246d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.132124] env[61868]: DEBUG oslo_vmware.api [None req-8b0db86f-df1c-40ac-889f-6d9f939a48bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 1036.132124] env[61868]: value = "task-1315930" [ 1036.132124] env[61868]: _type = "Task" [ 1036.132124] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.140820] env[61868]: DEBUG oslo_vmware.api [None req-8b0db86f-df1c-40ac-889f-6d9f939a48bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315930, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.353811] env[61868]: DEBUG oslo_vmware.api [None req-e85de296-a39d-423a-a0c2-369ae15a0786 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315915, 'name': ReconfigVM_Task, 'duration_secs': 5.762804} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.354186] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e85de296-a39d-423a-a0c2-369ae15a0786 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Releasing lock "a96cab5a-c053-4e45-96f6-2aba0a819110" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.354479] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-e85de296-a39d-423a-a0c2-369ae15a0786 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Reconfigured VM to detach interface {{(pid=61868) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1036.512422] env[61868]: DEBUG nova.compute.manager [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1036.535806] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8a8748ac-a070-45b9-b2c8-e2562a82a4d0 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "eeeec752-7e99-471d-9425-c4ccf4f6d0a5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.479s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.645201] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6659e422-d168-418e-bc02-73096ae733e6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.649635] env[61868]: DEBUG oslo_vmware.api [None req-8b0db86f-df1c-40ac-889f-6d9f939a48bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315930, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144655} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.650389] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b0db86f-df1c-40ac-889f-6d9f939a48bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1036.651017] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-8b0db86f-df1c-40ac-889f-6d9f939a48bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1036.651648] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-8b0db86f-df1c-40ac-889f-6d9f939a48bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1036.651648] env[61868]: INFO nova.compute.manager [None req-8b0db86f-df1c-40ac-889f-6d9f939a48bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1036.651901] env[61868]: DEBUG oslo.service.loopingcall [None req-8b0db86f-df1c-40ac-889f-6d9f939a48bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1036.655466] env[61868]: DEBUG nova.compute.manager [-] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1036.655466] env[61868]: DEBUG nova.network.neutron [-] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1036.657438] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa92241a-4ac4-4b46-93bf-b6f77e777c17 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.691772] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e19e6c55-9d55-4725-b651-8e8ed83a6f7b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.700500] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a587fadd-eb62-408d-8d6d-dc16057185a3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.715408] env[61868]: DEBUG nova.compute.provider_tree [None req-eda417b9-d960-4b01-bd7b-829d1de897c2 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1036.913509] env[61868]: DEBUG nova.compute.manager [req-2f1e5639-7bab-4d16-a96f-7a7910b659b6 req-7cc6eab3-5bf1-4014-be75-b9273da7feca service nova] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Received event network-vif-deleted-5736bca4-8f2f-485a-bc60-41a19416448c {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1036.913824] env[61868]: INFO nova.compute.manager [req-2f1e5639-7bab-4d16-a96f-7a7910b659b6 req-7cc6eab3-5bf1-4014-be75-b9273da7feca service nova] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Neutron deleted interface 5736bca4-8f2f-485a-bc60-41a19416448c; detaching it from the instance and deleting it from the info cache [ 1036.913824] env[61868]: DEBUG nova.network.neutron [req-2f1e5639-7bab-4d16-a96f-7a7910b659b6 req-7cc6eab3-5bf1-4014-be75-b9273da7feca service nova] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.030089] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.147161] env[61868]: DEBUG oslo_concurrency.lockutils [None req-84ba2995-5d5a-409e-9308-4bf7d312a3d8 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "e59a182d-97b2-454a-bc40-8afb0839324f" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.147424] env[61868]: DEBUG oslo_concurrency.lockutils [None req-84ba2995-5d5a-409e-9308-4bf7d312a3d8 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "e59a182d-97b2-454a-bc40-8afb0839324f" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.221827] env[61868]: DEBUG nova.scheduler.client.report [None req-eda417b9-d960-4b01-bd7b-829d1de897c2 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1037.390847] env[61868]: DEBUG nova.network.neutron [-] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.417042] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-98421fbc-cfa9-4067-8497-04299b4b5dba {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.427443] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4c17c71-edc3-4678-9696-4a878da13889 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.450781] env[61868]: DEBUG nova.compute.manager [req-2f1e5639-7bab-4d16-a96f-7a7910b659b6 req-7cc6eab3-5bf1-4014-be75-b9273da7feca service nova] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Detach interface failed, port_id=5736bca4-8f2f-485a-bc60-41a19416448c, reason: Instance cbeeb7b1-166b-49f7-88c5-6ed05c9dc759 could not be found. {{(pid=61868) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1037.650983] env[61868]: INFO nova.compute.manager [None req-84ba2995-5d5a-409e-9308-4bf7d312a3d8 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Detaching volume d18e4cb9-4034-4576-90e4-7458bce2ca23 [ 1037.685287] env[61868]: INFO nova.virt.block_device [None req-84ba2995-5d5a-409e-9308-4bf7d312a3d8 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Attempting to driver detach volume d18e4cb9-4034-4576-90e4-7458bce2ca23 from mountpoint /dev/sdb [ 1037.685539] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-84ba2995-5d5a-409e-9308-4bf7d312a3d8 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Volume detach. Driver type: vmdk {{(pid=61868) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1037.685731] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-84ba2995-5d5a-409e-9308-4bf7d312a3d8 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281596', 'volume_id': 'd18e4cb9-4034-4576-90e4-7458bce2ca23', 'name': 'volume-d18e4cb9-4034-4576-90e4-7458bce2ca23', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e59a182d-97b2-454a-bc40-8afb0839324f', 'attached_at': '', 'detached_at': '', 'volume_id': 'd18e4cb9-4034-4576-90e4-7458bce2ca23', 'serial': 'd18e4cb9-4034-4576-90e4-7458bce2ca23'} {{(pid=61868) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1037.686619] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-598d01dd-796d-44f3-a919-f3212ad03d73 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.708834] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b13d6d2-2a04-49d0-a608-750983add7ea {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.715519] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e162bd44-ac90-4e04-920d-89f7ecf7c256 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.735285] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eda417b9-d960-4b01-bd7b-829d1de897c2 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.731s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.737475] env[61868]: DEBUG oslo_concurrency.lockutils [None req-84e02b16-e99c-417d-8084-56b7f32bccb3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.631s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.737716] env[61868]: DEBUG nova.objects.instance [None req-84e02b16-e99c-417d-8084-56b7f32bccb3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lazy-loading 'resources' on Instance uuid 0316bfec-128e-41dc-ad3a-9ba3844b95fa {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1037.740581] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38f9385f-1a74-4c86-a671-017d384357c6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.743041] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e85de296-a39d-423a-a0c2-369ae15a0786 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "refresh_cache-a96cab5a-c053-4e45-96f6-2aba0a819110" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.743041] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e85de296-a39d-423a-a0c2-369ae15a0786 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquired lock "refresh_cache-a96cab5a-c053-4e45-96f6-2aba0a819110" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.743146] env[61868]: DEBUG nova.network.neutron [None req-e85de296-a39d-423a-a0c2-369ae15a0786 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1037.757293] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-84ba2995-5d5a-409e-9308-4bf7d312a3d8 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] The volume has not been displaced from its original location: [datastore2] volume-d18e4cb9-4034-4576-90e4-7458bce2ca23/volume-d18e4cb9-4034-4576-90e4-7458bce2ca23.vmdk. No consolidation needed. {{(pid=61868) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1037.763022] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-84ba2995-5d5a-409e-9308-4bf7d312a3d8 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Reconfiguring VM instance instance-0000004f to detach disk 2001 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1037.763686] env[61868]: INFO nova.scheduler.client.report [None req-eda417b9-d960-4b01-bd7b-829d1de897c2 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Deleted allocations for instance d9e90641-6765-4bda-ab07-3506898f06a9 [ 1037.767804] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-acb118ed-58b3-4c13-bbf8-4b25f8f1416b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.786504] env[61868]: DEBUG oslo_vmware.api [None req-84ba2995-5d5a-409e-9308-4bf7d312a3d8 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 1037.786504] env[61868]: value = "task-1315931" [ 1037.786504] env[61868]: _type = "Task" [ 1037.786504] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.794699] env[61868]: DEBUG oslo_vmware.api [None req-84ba2995-5d5a-409e-9308-4bf7d312a3d8 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315931, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.893696] env[61868]: INFO nova.compute.manager [-] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Took 1.24 seconds to deallocate network for instance. [ 1038.286074] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eda417b9-d960-4b01-bd7b-829d1de897c2 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "d9e90641-6765-4bda-ab07-3506898f06a9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.978s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.303355] env[61868]: DEBUG oslo_vmware.api [None req-84ba2995-5d5a-409e-9308-4bf7d312a3d8 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315931, 'name': ReconfigVM_Task, 'duration_secs': 0.257248} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.303640] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-84ba2995-5d5a-409e-9308-4bf7d312a3d8 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Reconfigured VM instance instance-0000004f to detach disk 2001 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1038.312835] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e123f61e-432c-406d-8176-2764ffce1f98 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.328922] env[61868]: DEBUG oslo_vmware.api [None req-84ba2995-5d5a-409e-9308-4bf7d312a3d8 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 1038.328922] env[61868]: value = "task-1315932" [ 1038.328922] env[61868]: _type = "Task" [ 1038.328922] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.338561] env[61868]: DEBUG oslo_vmware.api [None req-84ba2995-5d5a-409e-9308-4bf7d312a3d8 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315932, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.383403] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29180812-3dc4-4591-b1a0-337212dd0d03 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.392340] env[61868]: DEBUG nova.compute.manager [req-ab824c78-f8e6-41e0-a840-a6862fb1d941 req-3db73899-96d8-4db3-b568-4a6c6c5991ee service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Received event network-changed-cd9efdc6-e6f5-4449-b95d-d1fbec53a295 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1038.392463] env[61868]: DEBUG nova.compute.manager [req-ab824c78-f8e6-41e0-a840-a6862fb1d941 req-3db73899-96d8-4db3-b568-4a6c6c5991ee service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Refreshing instance network info cache due to event network-changed-cd9efdc6-e6f5-4449-b95d-d1fbec53a295. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1038.392699] env[61868]: DEBUG oslo_concurrency.lockutils [req-ab824c78-f8e6-41e0-a840-a6862fb1d941 req-3db73899-96d8-4db3-b568-4a6c6c5991ee service nova] Acquiring lock "refresh_cache-a96cab5a-c053-4e45-96f6-2aba0a819110" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1038.399918] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8b0db86f-df1c-40ac-889f-6d9f939a48bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.400940] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbf11eea-d04e-45e9-95c7-7acd53c69c09 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.434640] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dc973f8-a01c-473f-8ade-4fd49a692b7e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.442592] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb3013af-6fe5-45e8-a9c5-a7e4c52753a6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.456081] env[61868]: DEBUG nova.compute.provider_tree [None req-84e02b16-e99c-417d-8084-56b7f32bccb3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1038.510042] env[61868]: INFO nova.network.neutron [None req-e85de296-a39d-423a-a0c2-369ae15a0786 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Port 0bbbc97a-2c41-4d5d-a093-fca57a1d543b from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1038.510405] env[61868]: DEBUG nova.network.neutron [None req-e85de296-a39d-423a-a0c2-369ae15a0786 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Updating instance_info_cache with network_info: [{"id": "cd9efdc6-e6f5-4449-b95d-d1fbec53a295", "address": "fa:16:3e:bf:13:19", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd9efdc6-e6", "ovs_interfaceid": "cd9efdc6-e6f5-4449-b95d-d1fbec53a295", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.839100] env[61868]: DEBUG oslo_vmware.api [None req-84ba2995-5d5a-409e-9308-4bf7d312a3d8 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315932, 'name': ReconfigVM_Task, 'duration_secs': 0.134506} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.839410] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-84ba2995-5d5a-409e-9308-4bf7d312a3d8 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281596', 'volume_id': 'd18e4cb9-4034-4576-90e4-7458bce2ca23', 'name': 'volume-d18e4cb9-4034-4576-90e4-7458bce2ca23', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e59a182d-97b2-454a-bc40-8afb0839324f', 'attached_at': '', 'detached_at': '', 'volume_id': 'd18e4cb9-4034-4576-90e4-7458bce2ca23', 'serial': 'd18e4cb9-4034-4576-90e4-7458bce2ca23'} {{(pid=61868) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1038.962084] env[61868]: DEBUG nova.scheduler.client.report [None req-84e02b16-e99c-417d-8084-56b7f32bccb3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1039.013238] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e85de296-a39d-423a-a0c2-369ae15a0786 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Releasing lock "refresh_cache-a96cab5a-c053-4e45-96f6-2aba0a819110" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1039.015217] env[61868]: DEBUG oslo_concurrency.lockutils [req-ab824c78-f8e6-41e0-a840-a6862fb1d941 req-3db73899-96d8-4db3-b568-4a6c6c5991ee service nova] Acquired lock "refresh_cache-a96cab5a-c053-4e45-96f6-2aba0a819110" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.015407] env[61868]: DEBUG nova.network.neutron [req-ab824c78-f8e6-41e0-a840-a6862fb1d941 req-3db73899-96d8-4db3-b568-4a6c6c5991ee service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Refreshing network info cache for port cd9efdc6-e6f5-4449-b95d-d1fbec53a295 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1039.058294] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1fc3feee-4cb8-4a1f-8ced-ce06626c8481 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "interface-69a0b94b-2bf4-49ad-8fb0-ea945176e5eb-0bbbc97a-2c41-4d5d-a093-fca57a1d543b" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.058526] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1fc3feee-4cb8-4a1f-8ced-ce06626c8481 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "interface-69a0b94b-2bf4-49ad-8fb0-ea945176e5eb-0bbbc97a-2c41-4d5d-a093-fca57a1d543b" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.058865] env[61868]: DEBUG nova.objects.instance [None req-1fc3feee-4cb8-4a1f-8ced-ce06626c8481 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lazy-loading 'flavor' on Instance uuid 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1039.389101] env[61868]: DEBUG nova.objects.instance [None req-84ba2995-5d5a-409e-9308-4bf7d312a3d8 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lazy-loading 'flavor' on Instance uuid e59a182d-97b2-454a-bc40-8afb0839324f {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1039.424453] env[61868]: DEBUG oslo_concurrency.lockutils [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "3a94eb16-cce3-4765-bc61-b6776bdd5b63" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.425105] env[61868]: DEBUG oslo_concurrency.lockutils [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "3a94eb16-cce3-4765-bc61-b6776bdd5b63" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.466568] env[61868]: DEBUG oslo_concurrency.lockutils [None req-84e02b16-e99c-417d-8084-56b7f32bccb3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.729s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.468811] env[61868]: DEBUG oslo_concurrency.lockutils [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.164s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.469094] env[61868]: DEBUG oslo_concurrency.lockutils [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.470723] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.441s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.472156] env[61868]: INFO nova.compute.claims [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1039.489821] env[61868]: INFO nova.scheduler.client.report [None req-84e02b16-e99c-417d-8084-56b7f32bccb3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Deleted allocations for instance 0316bfec-128e-41dc-ad3a-9ba3844b95fa [ 1039.495114] env[61868]: INFO nova.scheduler.client.report [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Deleted allocations for instance 54dacd73-c760-49bd-9baf-ad2012a53264 [ 1039.519425] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e85de296-a39d-423a-a0c2-369ae15a0786 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "interface-a96cab5a-c053-4e45-96f6-2aba0a819110-0bbbc97a-2c41-4d5d-a093-fca57a1d543b" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.758s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.652187] env[61868]: DEBUG nova.objects.instance [None req-1fc3feee-4cb8-4a1f-8ced-ce06626c8481 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lazy-loading 'pci_requests' on Instance uuid 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1039.710791] env[61868]: DEBUG nova.network.neutron [req-ab824c78-f8e6-41e0-a840-a6862fb1d941 req-3db73899-96d8-4db3-b568-4a6c6c5991ee service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Updated VIF entry in instance network info cache for port cd9efdc6-e6f5-4449-b95d-d1fbec53a295. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1039.711189] env[61868]: DEBUG nova.network.neutron [req-ab824c78-f8e6-41e0-a840-a6862fb1d941 req-3db73899-96d8-4db3-b568-4a6c6c5991ee service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Updating instance_info_cache with network_info: [{"id": "cd9efdc6-e6f5-4449-b95d-d1fbec53a295", "address": "fa:16:3e:bf:13:19", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd9efdc6-e6", "ovs_interfaceid": "cd9efdc6-e6f5-4449-b95d-d1fbec53a295", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.927106] env[61868]: DEBUG nova.compute.manager [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1039.996947] env[61868]: DEBUG oslo_concurrency.lockutils [None req-84e02b16-e99c-417d-8084-56b7f32bccb3 tempest-ServerDiskConfigTestJSON-65177168 tempest-ServerDiskConfigTestJSON-65177168-project-member] Lock "0316bfec-128e-41dc-ad3a-9ba3844b95fa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.753s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.002445] env[61868]: DEBUG oslo_concurrency.lockutils [None req-51edf134-aa69-4258-a348-549cd2bc4322 tempest-DeleteServersTestJSON-777993345 tempest-DeleteServersTestJSON-777993345-project-member] Lock "54dacd73-c760-49bd-9baf-ad2012a53264" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.550s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.154584] env[61868]: DEBUG nova.objects.base [None req-1fc3feee-4cb8-4a1f-8ced-ce06626c8481 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Object Instance<69a0b94b-2bf4-49ad-8fb0-ea945176e5eb> lazy-loaded attributes: flavor,pci_requests {{(pid=61868) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1040.154824] env[61868]: DEBUG nova.network.neutron [None req-1fc3feee-4cb8-4a1f-8ced-ce06626c8481 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1040.214325] env[61868]: DEBUG oslo_concurrency.lockutils [req-ab824c78-f8e6-41e0-a840-a6862fb1d941 req-3db73899-96d8-4db3-b568-4a6c6c5991ee service nova] Releasing lock "refresh_cache-a96cab5a-c053-4e45-96f6-2aba0a819110" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1040.217698] env[61868]: DEBUG nova.policy [None req-1fc3feee-4cb8-4a1f-8ced-ce06626c8481 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c14b4ff01717495ca97c7f35f91c2995', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd7fc76299baf4a97b57139e5f1caa16e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 1040.397287] env[61868]: DEBUG oslo_concurrency.lockutils [None req-84ba2995-5d5a-409e-9308-4bf7d312a3d8 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "e59a182d-97b2-454a-bc40-8afb0839324f" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.250s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.421375] env[61868]: DEBUG nova.compute.manager [req-af81f8d7-2961-437b-b140-6b05ffdd8793 req-a0cd39f4-f135-406c-9bb2-7a6c91b04d28 service nova] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Received event network-changed-f8f4c521-03cd-4e54-99f1-761d9a515aa3 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1040.421721] env[61868]: DEBUG nova.compute.manager [req-af81f8d7-2961-437b-b140-6b05ffdd8793 req-a0cd39f4-f135-406c-9bb2-7a6c91b04d28 service nova] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Refreshing instance network info cache due to event network-changed-f8f4c521-03cd-4e54-99f1-761d9a515aa3. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1040.421791] env[61868]: DEBUG oslo_concurrency.lockutils [req-af81f8d7-2961-437b-b140-6b05ffdd8793 req-a0cd39f4-f135-406c-9bb2-7a6c91b04d28 service nova] Acquiring lock "refresh_cache-69a0b94b-2bf4-49ad-8fb0-ea945176e5eb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.421936] env[61868]: DEBUG oslo_concurrency.lockutils [req-af81f8d7-2961-437b-b140-6b05ffdd8793 req-a0cd39f4-f135-406c-9bb2-7a6c91b04d28 service nova] Acquired lock "refresh_cache-69a0b94b-2bf4-49ad-8fb0-ea945176e5eb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.422401] env[61868]: DEBUG nova.network.neutron [req-af81f8d7-2961-437b-b140-6b05ffdd8793 req-a0cd39f4-f135-406c-9bb2-7a6c91b04d28 service nova] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Refreshing network info cache for port f8f4c521-03cd-4e54-99f1-761d9a515aa3 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1040.446617] env[61868]: DEBUG oslo_concurrency.lockutils [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.555741] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37509267-786d-4370-95f1-d83f8fb03815 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.562476] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f31caad-e242-4cb0-bc4a-f72b32eef284 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.594540] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09b583f9-4702-43e0-a660-8269eeb1d003 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.601529] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0020f96b-1954-4896-b32c-46a4808fc818 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.605438] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9fb7bdcb-0c82-4801-8ae1-73a33bfc5086 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "e59a182d-97b2-454a-bc40-8afb0839324f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.605678] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9fb7bdcb-0c82-4801-8ae1-73a33bfc5086 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "e59a182d-97b2-454a-bc40-8afb0839324f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.605882] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9fb7bdcb-0c82-4801-8ae1-73a33bfc5086 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "e59a182d-97b2-454a-bc40-8afb0839324f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.607673] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9fb7bdcb-0c82-4801-8ae1-73a33bfc5086 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "e59a182d-97b2-454a-bc40-8afb0839324f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.607673] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9fb7bdcb-0c82-4801-8ae1-73a33bfc5086 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "e59a182d-97b2-454a-bc40-8afb0839324f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.608789] env[61868]: INFO nova.compute.manager [None req-9fb7bdcb-0c82-4801-8ae1-73a33bfc5086 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Terminating instance [ 1040.617656] env[61868]: DEBUG nova.compute.provider_tree [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1040.619254] env[61868]: DEBUG nova.compute.manager [None req-9fb7bdcb-0c82-4801-8ae1-73a33bfc5086 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1040.619451] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9fb7bdcb-0c82-4801-8ae1-73a33bfc5086 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1040.620211] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaf2695c-e48e-4efb-b0df-85ddd0b04b7a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.627022] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fb7bdcb-0c82-4801-8ae1-73a33bfc5086 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1040.627809] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1983505c-92fc-4301-9ca2-dad52958394e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.634229] env[61868]: DEBUG oslo_vmware.api [None req-9fb7bdcb-0c82-4801-8ae1-73a33bfc5086 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 1040.634229] env[61868]: value = "task-1315933" [ 1040.634229] env[61868]: _type = "Task" [ 1040.634229] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.642624] env[61868]: DEBUG oslo_vmware.api [None req-9fb7bdcb-0c82-4801-8ae1-73a33bfc5086 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315933, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.123734] env[61868]: DEBUG nova.scheduler.client.report [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1041.147076] env[61868]: DEBUG oslo_vmware.api [None req-9fb7bdcb-0c82-4801-8ae1-73a33bfc5086 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315933, 'name': PowerOffVM_Task, 'duration_secs': 0.182618} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.147778] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fb7bdcb-0c82-4801-8ae1-73a33bfc5086 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1041.148039] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9fb7bdcb-0c82-4801-8ae1-73a33bfc5086 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1041.148306] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ef349855-71c7-4c7e-aaab-0dd6a984c884 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.218015] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9fb7bdcb-0c82-4801-8ae1-73a33bfc5086 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1041.218259] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9fb7bdcb-0c82-4801-8ae1-73a33bfc5086 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Deleting contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1041.218510] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fb7bdcb-0c82-4801-8ae1-73a33bfc5086 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Deleting the datastore file [datastore2] e59a182d-97b2-454a-bc40-8afb0839324f {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1041.218708] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d0a6ec53-e30c-4334-bd0b-50a2fa48911e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.225082] env[61868]: DEBUG oslo_vmware.api [None req-9fb7bdcb-0c82-4801-8ae1-73a33bfc5086 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 1041.225082] env[61868]: value = "task-1315935" [ 1041.225082] env[61868]: _type = "Task" [ 1041.225082] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.235424] env[61868]: DEBUG oslo_vmware.api [None req-9fb7bdcb-0c82-4801-8ae1-73a33bfc5086 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315935, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.316697] env[61868]: DEBUG nova.network.neutron [req-af81f8d7-2961-437b-b140-6b05ffdd8793 req-a0cd39f4-f135-406c-9bb2-7a6c91b04d28 service nova] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Updated VIF entry in instance network info cache for port f8f4c521-03cd-4e54-99f1-761d9a515aa3. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1041.317106] env[61868]: DEBUG nova.network.neutron [req-af81f8d7-2961-437b-b140-6b05ffdd8793 req-a0cd39f4-f135-406c-9bb2-7a6c91b04d28 service nova] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Updating instance_info_cache with network_info: [{"id": "f8f4c521-03cd-4e54-99f1-761d9a515aa3", "address": "fa:16:3e:3c:dd:65", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8f4c521-03", "ovs_interfaceid": "f8f4c521-03cd-4e54-99f1-761d9a515aa3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.628655] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.158s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.629281] env[61868]: DEBUG nova.compute.manager [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1041.631719] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8b0db86f-df1c-40ac-889f-6d9f939a48bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.232s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.635105] env[61868]: DEBUG nova.objects.instance [None req-8b0db86f-df1c-40ac-889f-6d9f939a48bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lazy-loading 'resources' on Instance uuid cbeeb7b1-166b-49f7-88c5-6ed05c9dc759 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1041.713099] env[61868]: DEBUG nova.compute.manager [req-22d2baa0-bf83-40f3-b370-e5662ebfcd90 req-18a309a8-37b0-4cfb-88b6-183e9accbc1c service nova] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Received event network-vif-plugged-0bbbc97a-2c41-4d5d-a093-fca57a1d543b {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1041.713356] env[61868]: DEBUG oslo_concurrency.lockutils [req-22d2baa0-bf83-40f3-b370-e5662ebfcd90 req-18a309a8-37b0-4cfb-88b6-183e9accbc1c service nova] Acquiring lock "69a0b94b-2bf4-49ad-8fb0-ea945176e5eb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.713529] env[61868]: DEBUG oslo_concurrency.lockutils [req-22d2baa0-bf83-40f3-b370-e5662ebfcd90 req-18a309a8-37b0-4cfb-88b6-183e9accbc1c service nova] Lock "69a0b94b-2bf4-49ad-8fb0-ea945176e5eb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.713698] env[61868]: DEBUG oslo_concurrency.lockutils [req-22d2baa0-bf83-40f3-b370-e5662ebfcd90 req-18a309a8-37b0-4cfb-88b6-183e9accbc1c service nova] Lock "69a0b94b-2bf4-49ad-8fb0-ea945176e5eb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.713871] env[61868]: DEBUG nova.compute.manager [req-22d2baa0-bf83-40f3-b370-e5662ebfcd90 req-18a309a8-37b0-4cfb-88b6-183e9accbc1c service nova] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] No waiting events found dispatching network-vif-plugged-0bbbc97a-2c41-4d5d-a093-fca57a1d543b {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1041.714232] env[61868]: WARNING nova.compute.manager [req-22d2baa0-bf83-40f3-b370-e5662ebfcd90 req-18a309a8-37b0-4cfb-88b6-183e9accbc1c service nova] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Received unexpected event network-vif-plugged-0bbbc97a-2c41-4d5d-a093-fca57a1d543b for instance with vm_state active and task_state None. [ 1041.736171] env[61868]: DEBUG oslo_vmware.api [None req-9fb7bdcb-0c82-4801-8ae1-73a33bfc5086 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315935, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138417} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.736171] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fb7bdcb-0c82-4801-8ae1-73a33bfc5086 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1041.736171] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9fb7bdcb-0c82-4801-8ae1-73a33bfc5086 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Deleted contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1041.736401] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9fb7bdcb-0c82-4801-8ae1-73a33bfc5086 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1041.736572] env[61868]: INFO nova.compute.manager [None req-9fb7bdcb-0c82-4801-8ae1-73a33bfc5086 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1041.736816] env[61868]: DEBUG oslo.service.loopingcall [None req-9fb7bdcb-0c82-4801-8ae1-73a33bfc5086 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1041.737014] env[61868]: DEBUG nova.compute.manager [-] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1041.737113] env[61868]: DEBUG nova.network.neutron [-] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1041.819555] env[61868]: DEBUG oslo_concurrency.lockutils [req-af81f8d7-2961-437b-b140-6b05ffdd8793 req-a0cd39f4-f135-406c-9bb2-7a6c91b04d28 service nova] Releasing lock "refresh_cache-69a0b94b-2bf4-49ad-8fb0-ea945176e5eb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.067508] env[61868]: DEBUG nova.network.neutron [None req-1fc3feee-4cb8-4a1f-8ced-ce06626c8481 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Successfully updated port: 0bbbc97a-2c41-4d5d-a093-fca57a1d543b {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1042.135664] env[61868]: DEBUG nova.compute.utils [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1042.137504] env[61868]: DEBUG nova.compute.manager [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1042.138223] env[61868]: DEBUG nova.network.neutron [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1042.208574] env[61868]: DEBUG nova.policy [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0b615dd3d1064706bea31b52fbfa9a61', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5658848e1b0e42929c04a4a8de40a291', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 1042.250842] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51fcedee-5955-43c8-87dc-ddf8854ebbb1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.259825] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f40cb27-a5eb-4bb3-aa53-04486b2ba704 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.294525] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89257b8e-297d-4300-aea2-68f3fe84853f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.302659] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5fe6441-70a7-4e71-8d70-40f4a39eb495 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.316937] env[61868]: DEBUG nova.compute.provider_tree [None req-8b0db86f-df1c-40ac-889f-6d9f939a48bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1042.489620] env[61868]: DEBUG nova.compute.manager [req-5f74beed-12af-4154-8f83-7be3219ea2ab req-0cbb9bb9-4849-4b43-bc36-05bf2f457d2b service nova] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Received event network-vif-deleted-cb19d64d-19c9-4d48-8d70-00b3440d9127 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1042.490196] env[61868]: INFO nova.compute.manager [req-5f74beed-12af-4154-8f83-7be3219ea2ab req-0cbb9bb9-4849-4b43-bc36-05bf2f457d2b service nova] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Neutron deleted interface cb19d64d-19c9-4d48-8d70-00b3440d9127; detaching it from the instance and deleting it from the info cache [ 1042.490554] env[61868]: DEBUG nova.network.neutron [req-5f74beed-12af-4154-8f83-7be3219ea2ab req-0cbb9bb9-4849-4b43-bc36-05bf2f457d2b service nova] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.571351] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1fc3feee-4cb8-4a1f-8ced-ce06626c8481 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "refresh_cache-69a0b94b-2bf4-49ad-8fb0-ea945176e5eb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.572104] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1fc3feee-4cb8-4a1f-8ced-ce06626c8481 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquired lock "refresh_cache-69a0b94b-2bf4-49ad-8fb0-ea945176e5eb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.572104] env[61868]: DEBUG nova.network.neutron [None req-1fc3feee-4cb8-4a1f-8ced-ce06626c8481 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1042.648096] env[61868]: DEBUG nova.compute.manager [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1042.677723] env[61868]: DEBUG nova.network.neutron [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Successfully created port: 89658ae1-cd68-4ab5-9ddd-7de554905b00 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1042.820436] env[61868]: DEBUG nova.scheduler.client.report [None req-8b0db86f-df1c-40ac-889f-6d9f939a48bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1042.920478] env[61868]: DEBUG nova.network.neutron [-] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.993837] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4cc6e285-0f89-4cda-94f0-55d89672e735 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.003648] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6486ba6c-17ae-488e-b15d-8426b4031078 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.028515] env[61868]: DEBUG nova.compute.manager [req-5f74beed-12af-4154-8f83-7be3219ea2ab req-0cbb9bb9-4849-4b43-bc36-05bf2f457d2b service nova] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Detach interface failed, port_id=cb19d64d-19c9-4d48-8d70-00b3440d9127, reason: Instance e59a182d-97b2-454a-bc40-8afb0839324f could not be found. {{(pid=61868) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1043.107959] env[61868]: WARNING nova.network.neutron [None req-1fc3feee-4cb8-4a1f-8ced-ce06626c8481 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] d4f100cd-8de6-4459-a7a6-bc1895b8b14f already exists in list: networks containing: ['d4f100cd-8de6-4459-a7a6-bc1895b8b14f']. ignoring it [ 1043.325874] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8b0db86f-df1c-40ac-889f-6d9f939a48bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.694s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.328575] env[61868]: DEBUG oslo_concurrency.lockutils [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.882s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.330569] env[61868]: INFO nova.compute.claims [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1043.357800] env[61868]: INFO nova.scheduler.client.report [None req-8b0db86f-df1c-40ac-889f-6d9f939a48bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Deleted allocations for instance cbeeb7b1-166b-49f7-88c5-6ed05c9dc759 [ 1043.423785] env[61868]: INFO nova.compute.manager [-] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Took 1.69 seconds to deallocate network for instance. [ 1043.496762] env[61868]: DEBUG nova.network.neutron [None req-1fc3feee-4cb8-4a1f-8ced-ce06626c8481 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Updating instance_info_cache with network_info: [{"id": "f8f4c521-03cd-4e54-99f1-761d9a515aa3", "address": "fa:16:3e:3c:dd:65", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8f4c521-03", "ovs_interfaceid": "f8f4c521-03cd-4e54-99f1-761d9a515aa3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0bbbc97a-2c41-4d5d-a093-fca57a1d543b", "address": "fa:16:3e:b0:b6:94", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0bbbc97a-2c", "ovs_interfaceid": "0bbbc97a-2c41-4d5d-a093-fca57a1d543b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.658602] env[61868]: DEBUG nova.compute.manager [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1043.684436] env[61868]: DEBUG nova.virt.hardware [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1043.684681] env[61868]: DEBUG nova.virt.hardware [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1043.684847] env[61868]: DEBUG nova.virt.hardware [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1043.685053] env[61868]: DEBUG nova.virt.hardware [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1043.685215] env[61868]: DEBUG nova.virt.hardware [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1043.685366] env[61868]: DEBUG nova.virt.hardware [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1043.685580] env[61868]: DEBUG nova.virt.hardware [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1043.685745] env[61868]: DEBUG nova.virt.hardware [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1043.685928] env[61868]: DEBUG nova.virt.hardware [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1043.686098] env[61868]: DEBUG nova.virt.hardware [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1043.686283] env[61868]: DEBUG nova.virt.hardware [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1043.687160] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c4fef38-7af6-4754-8663-1ee06cc7e107 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.694945] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a9aad68-702a-4077-8dde-d59af9677f79 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.746112] env[61868]: DEBUG nova.compute.manager [req-9813349c-a117-48ca-ae6e-520a1fbcdbaf req-687b924a-2c6b-4da6-b348-c7d90cf865ea service nova] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Received event network-changed-0bbbc97a-2c41-4d5d-a093-fca57a1d543b {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1043.746360] env[61868]: DEBUG nova.compute.manager [req-9813349c-a117-48ca-ae6e-520a1fbcdbaf req-687b924a-2c6b-4da6-b348-c7d90cf865ea service nova] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Refreshing instance network info cache due to event network-changed-0bbbc97a-2c41-4d5d-a093-fca57a1d543b. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1043.746529] env[61868]: DEBUG oslo_concurrency.lockutils [req-9813349c-a117-48ca-ae6e-520a1fbcdbaf req-687b924a-2c6b-4da6-b348-c7d90cf865ea service nova] Acquiring lock "refresh_cache-69a0b94b-2bf4-49ad-8fb0-ea945176e5eb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.871704] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8b0db86f-df1c-40ac-889f-6d9f939a48bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "cbeeb7b1-166b-49f7-88c5-6ed05c9dc759" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.352s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.929810] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9fb7bdcb-0c82-4801-8ae1-73a33bfc5086 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.000737] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1fc3feee-4cb8-4a1f-8ced-ce06626c8481 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Releasing lock "refresh_cache-69a0b94b-2bf4-49ad-8fb0-ea945176e5eb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.001075] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1fc3feee-4cb8-4a1f-8ced-ce06626c8481 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "69a0b94b-2bf4-49ad-8fb0-ea945176e5eb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.001267] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1fc3feee-4cb8-4a1f-8ced-ce06626c8481 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquired lock "69a0b94b-2bf4-49ad-8fb0-ea945176e5eb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.001577] env[61868]: DEBUG oslo_concurrency.lockutils [req-9813349c-a117-48ca-ae6e-520a1fbcdbaf req-687b924a-2c6b-4da6-b348-c7d90cf865ea service nova] Acquired lock "refresh_cache-69a0b94b-2bf4-49ad-8fb0-ea945176e5eb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.001772] env[61868]: DEBUG nova.network.neutron [req-9813349c-a117-48ca-ae6e-520a1fbcdbaf req-687b924a-2c6b-4da6-b348-c7d90cf865ea service nova] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Refreshing network info cache for port 0bbbc97a-2c41-4d5d-a093-fca57a1d543b {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1044.004943] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4156de4d-aaf1-4f12-bdf4-8268d0f7576a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.028224] env[61868]: DEBUG nova.virt.hardware [None req-1fc3feee-4cb8-4a1f-8ced-ce06626c8481 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1044.028610] env[61868]: DEBUG nova.virt.hardware [None req-1fc3feee-4cb8-4a1f-8ced-ce06626c8481 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1044.028790] env[61868]: DEBUG nova.virt.hardware [None req-1fc3feee-4cb8-4a1f-8ced-ce06626c8481 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1044.028980] env[61868]: DEBUG nova.virt.hardware [None req-1fc3feee-4cb8-4a1f-8ced-ce06626c8481 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1044.029147] env[61868]: DEBUG nova.virt.hardware [None req-1fc3feee-4cb8-4a1f-8ced-ce06626c8481 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1044.029300] env[61868]: DEBUG nova.virt.hardware [None req-1fc3feee-4cb8-4a1f-8ced-ce06626c8481 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1044.029510] env[61868]: DEBUG nova.virt.hardware [None req-1fc3feee-4cb8-4a1f-8ced-ce06626c8481 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1044.029745] env[61868]: DEBUG nova.virt.hardware [None req-1fc3feee-4cb8-4a1f-8ced-ce06626c8481 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1044.029960] env[61868]: DEBUG nova.virt.hardware [None req-1fc3feee-4cb8-4a1f-8ced-ce06626c8481 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1044.030287] env[61868]: DEBUG nova.virt.hardware [None req-1fc3feee-4cb8-4a1f-8ced-ce06626c8481 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1044.030388] env[61868]: DEBUG nova.virt.hardware [None req-1fc3feee-4cb8-4a1f-8ced-ce06626c8481 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1044.038352] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1fc3feee-4cb8-4a1f-8ced-ce06626c8481 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Reconfiguring VM to attach interface {{(pid=61868) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1044.038914] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-039984a4-8095-4950-b7da-07cb6c4194fe {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.057902] env[61868]: DEBUG oslo_vmware.api [None req-1fc3feee-4cb8-4a1f-8ced-ce06626c8481 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for the task: (returnval){ [ 1044.057902] env[61868]: value = "task-1315936" [ 1044.057902] env[61868]: _type = "Task" [ 1044.057902] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.066668] env[61868]: DEBUG oslo_vmware.api [None req-1fc3feee-4cb8-4a1f-8ced-ce06626c8481 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315936, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.371718] env[61868]: DEBUG nova.network.neutron [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Successfully updated port: 89658ae1-cd68-4ab5-9ddd-7de554905b00 {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1044.453112] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1dcf5f0-7928-4f9a-a0a3-7846326d9569 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.461288] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdf5a09b-2ad3-4ecf-ba09-27f93dff3c63 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.500541] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7255382-7c24-4255-bd38-7a3a01b97b8d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.509623] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74cf364e-0c16-40b8-ae50-546aa6d6b687 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.526892] env[61868]: DEBUG nova.compute.provider_tree [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1044.569229] env[61868]: DEBUG oslo_vmware.api [None req-1fc3feee-4cb8-4a1f-8ced-ce06626c8481 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315936, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.857222] env[61868]: DEBUG nova.network.neutron [req-9813349c-a117-48ca-ae6e-520a1fbcdbaf req-687b924a-2c6b-4da6-b348-c7d90cf865ea service nova] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Updated VIF entry in instance network info cache for port 0bbbc97a-2c41-4d5d-a093-fca57a1d543b. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1044.857676] env[61868]: DEBUG nova.network.neutron [req-9813349c-a117-48ca-ae6e-520a1fbcdbaf req-687b924a-2c6b-4da6-b348-c7d90cf865ea service nova] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Updating instance_info_cache with network_info: [{"id": "f8f4c521-03cd-4e54-99f1-761d9a515aa3", "address": "fa:16:3e:3c:dd:65", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8f4c521-03", "ovs_interfaceid": "f8f4c521-03cd-4e54-99f1-761d9a515aa3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0bbbc97a-2c41-4d5d-a093-fca57a1d543b", "address": "fa:16:3e:b0:b6:94", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0bbbc97a-2c", "ovs_interfaceid": "0bbbc97a-2c41-4d5d-a093-fca57a1d543b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.875033] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "refresh_cache-7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.875222] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquired lock "refresh_cache-7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.875382] env[61868]: DEBUG nova.network.neutron [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1045.030126] env[61868]: DEBUG nova.scheduler.client.report [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1045.069152] env[61868]: DEBUG oslo_vmware.api [None req-1fc3feee-4cb8-4a1f-8ced-ce06626c8481 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315936, 'name': ReconfigVM_Task, 'duration_secs': 0.568422} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.069656] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1fc3feee-4cb8-4a1f-8ced-ce06626c8481 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Releasing lock "69a0b94b-2bf4-49ad-8fb0-ea945176e5eb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.069872] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1fc3feee-4cb8-4a1f-8ced-ce06626c8481 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Reconfigured VM to attach interface {{(pid=61868) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1045.363191] env[61868]: DEBUG oslo_concurrency.lockutils [req-9813349c-a117-48ca-ae6e-520a1fbcdbaf req-687b924a-2c6b-4da6-b348-c7d90cf865ea service nova] Releasing lock "refresh_cache-69a0b94b-2bf4-49ad-8fb0-ea945176e5eb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.415238] env[61868]: DEBUG nova.network.neutron [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1045.534869] env[61868]: DEBUG oslo_concurrency.lockutils [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.206s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.537217] env[61868]: DEBUG nova.compute.manager [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1045.538181] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9fb7bdcb-0c82-4801-8ae1-73a33bfc5086 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.608s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.538379] env[61868]: DEBUG nova.objects.instance [None req-9fb7bdcb-0c82-4801-8ae1-73a33bfc5086 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lazy-loading 'resources' on Instance uuid e59a182d-97b2-454a-bc40-8afb0839324f {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1045.574903] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1fc3feee-4cb8-4a1f-8ced-ce06626c8481 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "interface-69a0b94b-2bf4-49ad-8fb0-ea945176e5eb-0bbbc97a-2c41-4d5d-a093-fca57a1d543b" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.516s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.601028] env[61868]: DEBUG nova.network.neutron [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Updating instance_info_cache with network_info: [{"id": "89658ae1-cd68-4ab5-9ddd-7de554905b00", "address": "fa:16:3e:b8:82:d7", "network": {"id": "f2e3f911-d826-43ad-911d-9f6985db7ee5", "bridge": "br-int", "label": "tempest-ImagesTestJSON-485234144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5658848e1b0e42929c04a4a8de40a291", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap89658ae1-cd", "ovs_interfaceid": "89658ae1-cd68-4ab5-9ddd-7de554905b00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.031434] env[61868]: DEBUG nova.compute.manager [req-28d34f2e-02bf-4d23-ac35-55b47dc19cdf req-77f3476c-dc61-4883-bd06-8a4d0d83d6cf service nova] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Received event network-vif-plugged-89658ae1-cd68-4ab5-9ddd-7de554905b00 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1046.031737] env[61868]: DEBUG oslo_concurrency.lockutils [req-28d34f2e-02bf-4d23-ac35-55b47dc19cdf req-77f3476c-dc61-4883-bd06-8a4d0d83d6cf service nova] Acquiring lock "7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.031892] env[61868]: DEBUG oslo_concurrency.lockutils [req-28d34f2e-02bf-4d23-ac35-55b47dc19cdf req-77f3476c-dc61-4883-bd06-8a4d0d83d6cf service nova] Lock "7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.032073] env[61868]: DEBUG oslo_concurrency.lockutils [req-28d34f2e-02bf-4d23-ac35-55b47dc19cdf req-77f3476c-dc61-4883-bd06-8a4d0d83d6cf service nova] Lock "7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.032247] env[61868]: DEBUG nova.compute.manager [req-28d34f2e-02bf-4d23-ac35-55b47dc19cdf req-77f3476c-dc61-4883-bd06-8a4d0d83d6cf service nova] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] No waiting events found dispatching network-vif-plugged-89658ae1-cd68-4ab5-9ddd-7de554905b00 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1046.032417] env[61868]: WARNING nova.compute.manager [req-28d34f2e-02bf-4d23-ac35-55b47dc19cdf req-77f3476c-dc61-4883-bd06-8a4d0d83d6cf service nova] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Received unexpected event network-vif-plugged-89658ae1-cd68-4ab5-9ddd-7de554905b00 for instance with vm_state building and task_state spawning. [ 1046.032583] env[61868]: DEBUG nova.compute.manager [req-28d34f2e-02bf-4d23-ac35-55b47dc19cdf req-77f3476c-dc61-4883-bd06-8a4d0d83d6cf service nova] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Received event network-changed-89658ae1-cd68-4ab5-9ddd-7de554905b00 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1046.032898] env[61868]: DEBUG nova.compute.manager [req-28d34f2e-02bf-4d23-ac35-55b47dc19cdf req-77f3476c-dc61-4883-bd06-8a4d0d83d6cf service nova] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Refreshing instance network info cache due to event network-changed-89658ae1-cd68-4ab5-9ddd-7de554905b00. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1046.033040] env[61868]: DEBUG oslo_concurrency.lockutils [req-28d34f2e-02bf-4d23-ac35-55b47dc19cdf req-77f3476c-dc61-4883-bd06-8a4d0d83d6cf service nova] Acquiring lock "refresh_cache-7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.040905] env[61868]: DEBUG nova.compute.utils [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1046.045324] env[61868]: DEBUG nova.compute.manager [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1046.046253] env[61868]: DEBUG nova.network.neutron [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1046.085781] env[61868]: DEBUG nova.policy [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'daad105e7edf4fb0ae0b2e685bfd92e9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b520c3ac58074e8d9b0bfafb817244a8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 1046.103713] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Releasing lock "refresh_cache-7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.103713] env[61868]: DEBUG nova.compute.manager [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Instance network_info: |[{"id": "89658ae1-cd68-4ab5-9ddd-7de554905b00", "address": "fa:16:3e:b8:82:d7", "network": {"id": "f2e3f911-d826-43ad-911d-9f6985db7ee5", "bridge": "br-int", "label": "tempest-ImagesTestJSON-485234144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5658848e1b0e42929c04a4a8de40a291", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap89658ae1-cd", "ovs_interfaceid": "89658ae1-cd68-4ab5-9ddd-7de554905b00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1046.106383] env[61868]: DEBUG oslo_concurrency.lockutils [req-28d34f2e-02bf-4d23-ac35-55b47dc19cdf req-77f3476c-dc61-4883-bd06-8a4d0d83d6cf service nova] Acquired lock "refresh_cache-7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.106581] env[61868]: DEBUG nova.network.neutron [req-28d34f2e-02bf-4d23-ac35-55b47dc19cdf req-77f3476c-dc61-4883-bd06-8a4d0d83d6cf service nova] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Refreshing network info cache for port 89658ae1-cd68-4ab5-9ddd-7de554905b00 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1046.107800] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b8:82:d7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f1b507ed-cd2d-4c09-9d96-c47bde6a7774', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '89658ae1-cd68-4ab5-9ddd-7de554905b00', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1046.115236] env[61868]: DEBUG oslo.service.loopingcall [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1046.116680] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1046.116680] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-59322f1e-ca14-4c4d-a425-0da3bc47db78 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.138911] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1046.138911] env[61868]: value = "task-1315937" [ 1046.138911] env[61868]: _type = "Task" [ 1046.138911] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.147200] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315937, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.160521] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f455549-ce2e-41f9-96c3-95ebbdd9dc7b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.167555] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b852e9cd-f822-4ec4-88f9-5365c8b98a78 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.196690] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6b0ba0d-f218-48d1-8e63-9c5adbb9713d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.204706] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-190990ab-07b6-4347-81c6-41227f7f98d6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.219931] env[61868]: DEBUG nova.compute.provider_tree [None req-9fb7bdcb-0c82-4801-8ae1-73a33bfc5086 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1046.351646] env[61868]: DEBUG nova.network.neutron [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Successfully created port: 32232776-6205-48a9-abcc-b24add5b69a1 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1046.441268] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Acquiring lock "1cb04c98-f725-4b1f-933e-ded580d340fb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.441511] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Lock "1cb04c98-f725-4b1f-933e-ded580d340fb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.549697] env[61868]: DEBUG nova.compute.manager [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1046.650159] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315937, 'name': CreateVM_Task, 'duration_secs': 0.300975} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.650378] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1046.651229] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.651574] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.652040] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1046.652441] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-97fef550-2a21-44e8-972a-4c5d1f53210e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.657850] env[61868]: DEBUG oslo_vmware.api [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 1046.657850] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52fbdd5b-ce2d-2d28-ab41-9233b393260f" [ 1046.657850] env[61868]: _type = "Task" [ 1046.657850] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.668278] env[61868]: DEBUG oslo_vmware.api [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52fbdd5b-ce2d-2d28-ab41-9233b393260f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.727026] env[61868]: DEBUG nova.scheduler.client.report [None req-9fb7bdcb-0c82-4801-8ae1-73a33bfc5086 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1046.867853] env[61868]: DEBUG nova.network.neutron [req-28d34f2e-02bf-4d23-ac35-55b47dc19cdf req-77f3476c-dc61-4883-bd06-8a4d0d83d6cf service nova] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Updated VIF entry in instance network info cache for port 89658ae1-cd68-4ab5-9ddd-7de554905b00. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1046.868447] env[61868]: DEBUG nova.network.neutron [req-28d34f2e-02bf-4d23-ac35-55b47dc19cdf req-77f3476c-dc61-4883-bd06-8a4d0d83d6cf service nova] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Updating instance_info_cache with network_info: [{"id": "89658ae1-cd68-4ab5-9ddd-7de554905b00", "address": "fa:16:3e:b8:82:d7", "network": {"id": "f2e3f911-d826-43ad-911d-9f6985db7ee5", "bridge": "br-int", "label": "tempest-ImagesTestJSON-485234144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5658848e1b0e42929c04a4a8de40a291", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap89658ae1-cd", "ovs_interfaceid": "89658ae1-cd68-4ab5-9ddd-7de554905b00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.944093] env[61868]: DEBUG nova.compute.manager [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1047.173956] env[61868]: DEBUG oslo_vmware.api [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52fbdd5b-ce2d-2d28-ab41-9233b393260f, 'name': SearchDatastore_Task, 'duration_secs': 0.039669} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.174347] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.174640] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1047.174937] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.175161] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.175425] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1047.175764] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4cdb9e11-9e22-4d54-a22b-f2aefae20f46 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.185616] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1047.185796] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1047.186559] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c027e807-8a36-41b5-b807-fdefa4cf90c3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.192224] env[61868]: DEBUG oslo_vmware.api [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 1047.192224] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52f433bd-7068-4333-1c12-5dc473107e8c" [ 1047.192224] env[61868]: _type = "Task" [ 1047.192224] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.200319] env[61868]: DEBUG oslo_vmware.api [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52f433bd-7068-4333-1c12-5dc473107e8c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.233135] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9fb7bdcb-0c82-4801-8ae1-73a33bfc5086 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.695s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.257273] env[61868]: INFO nova.scheduler.client.report [None req-9fb7bdcb-0c82-4801-8ae1-73a33bfc5086 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Deleted allocations for instance e59a182d-97b2-454a-bc40-8afb0839324f [ 1047.316268] env[61868]: DEBUG oslo_concurrency.lockutils [None req-fd7d4ac0-445f-48d4-98fe-3fdc7b61e43c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "interface-69a0b94b-2bf4-49ad-8fb0-ea945176e5eb-0bbbc97a-2c41-4d5d-a093-fca57a1d543b" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.316567] env[61868]: DEBUG oslo_concurrency.lockutils [None req-fd7d4ac0-445f-48d4-98fe-3fdc7b61e43c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "interface-69a0b94b-2bf4-49ad-8fb0-ea945176e5eb-0bbbc97a-2c41-4d5d-a093-fca57a1d543b" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.372229] env[61868]: DEBUG oslo_concurrency.lockutils [req-28d34f2e-02bf-4d23-ac35-55b47dc19cdf req-77f3476c-dc61-4883-bd06-8a4d0d83d6cf service nova] Releasing lock "refresh_cache-7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.464983] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.465142] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.466676] env[61868]: INFO nova.compute.claims [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1047.563074] env[61868]: DEBUG nova.compute.manager [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1047.589211] env[61868]: DEBUG nova.virt.hardware [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1047.589466] env[61868]: DEBUG nova.virt.hardware [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1047.589628] env[61868]: DEBUG nova.virt.hardware [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1047.589815] env[61868]: DEBUG nova.virt.hardware [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1047.589967] env[61868]: DEBUG nova.virt.hardware [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1047.590143] env[61868]: DEBUG nova.virt.hardware [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1047.590353] env[61868]: DEBUG nova.virt.hardware [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1047.590517] env[61868]: DEBUG nova.virt.hardware [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1047.590685] env[61868]: DEBUG nova.virt.hardware [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1047.590864] env[61868]: DEBUG nova.virt.hardware [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1047.591081] env[61868]: DEBUG nova.virt.hardware [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1047.591952] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ae7e73c-7b7c-4007-8338-d5655409f77d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.600327] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4880169-fe5b-4a3e-835e-5dc86b4a8215 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.702706] env[61868]: DEBUG oslo_vmware.api [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52f433bd-7068-4333-1c12-5dc473107e8c, 'name': SearchDatastore_Task, 'duration_secs': 0.02412} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.703702] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e699dc7-1085-42cf-9cf9-defee9e84bea {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.708943] env[61868]: DEBUG oslo_vmware.api [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 1047.708943] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52611103-60f8-e6ce-5401-80d082752a7a" [ 1047.708943] env[61868]: _type = "Task" [ 1047.708943] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.717054] env[61868]: DEBUG oslo_vmware.api [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52611103-60f8-e6ce-5401-80d082752a7a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.765887] env[61868]: DEBUG oslo_concurrency.lockutils [None req-9fb7bdcb-0c82-4801-8ae1-73a33bfc5086 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "e59a182d-97b2-454a-bc40-8afb0839324f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.160s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.821172] env[61868]: DEBUG oslo_concurrency.lockutils [None req-fd7d4ac0-445f-48d4-98fe-3fdc7b61e43c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "69a0b94b-2bf4-49ad-8fb0-ea945176e5eb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.821484] env[61868]: DEBUG oslo_concurrency.lockutils [None req-fd7d4ac0-445f-48d4-98fe-3fdc7b61e43c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquired lock "69a0b94b-2bf4-49ad-8fb0-ea945176e5eb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.822285] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a3499b3-0a24-48f2-bfbe-9b265588f828 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.840559] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be93a56e-a2be-46ce-883c-6995158fa4e3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.867602] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-fd7d4ac0-445f-48d4-98fe-3fdc7b61e43c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Reconfiguring VM to detach interface {{(pid=61868) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1047.867956] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5c48a4c5-eea2-4a57-9e13-40404afad700 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.883163] env[61868]: DEBUG nova.network.neutron [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Successfully updated port: 32232776-6205-48a9-abcc-b24add5b69a1 {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1047.890957] env[61868]: DEBUG oslo_vmware.api [None req-fd7d4ac0-445f-48d4-98fe-3fdc7b61e43c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for the task: (returnval){ [ 1047.890957] env[61868]: value = "task-1315938" [ 1047.890957] env[61868]: _type = "Task" [ 1047.890957] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.900498] env[61868]: DEBUG oslo_vmware.api [None req-fd7d4ac0-445f-48d4-98fe-3fdc7b61e43c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315938, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.058950] env[61868]: DEBUG nova.compute.manager [req-687e3c80-1a1d-4229-ad10-07e27f131b5f req-74ac6fdc-3247-4f3d-9627-8da14089c8bf service nova] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Received event network-vif-plugged-32232776-6205-48a9-abcc-b24add5b69a1 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1048.059098] env[61868]: DEBUG oslo_concurrency.lockutils [req-687e3c80-1a1d-4229-ad10-07e27f131b5f req-74ac6fdc-3247-4f3d-9627-8da14089c8bf service nova] Acquiring lock "3a94eb16-cce3-4765-bc61-b6776bdd5b63-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.059297] env[61868]: DEBUG oslo_concurrency.lockutils [req-687e3c80-1a1d-4229-ad10-07e27f131b5f req-74ac6fdc-3247-4f3d-9627-8da14089c8bf service nova] Lock "3a94eb16-cce3-4765-bc61-b6776bdd5b63-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.059496] env[61868]: DEBUG oslo_concurrency.lockutils [req-687e3c80-1a1d-4229-ad10-07e27f131b5f req-74ac6fdc-3247-4f3d-9627-8da14089c8bf service nova] Lock "3a94eb16-cce3-4765-bc61-b6776bdd5b63-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.059635] env[61868]: DEBUG nova.compute.manager [req-687e3c80-1a1d-4229-ad10-07e27f131b5f req-74ac6fdc-3247-4f3d-9627-8da14089c8bf service nova] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] No waiting events found dispatching network-vif-plugged-32232776-6205-48a9-abcc-b24add5b69a1 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1048.059802] env[61868]: WARNING nova.compute.manager [req-687e3c80-1a1d-4229-ad10-07e27f131b5f req-74ac6fdc-3247-4f3d-9627-8da14089c8bf service nova] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Received unexpected event network-vif-plugged-32232776-6205-48a9-abcc-b24add5b69a1 for instance with vm_state building and task_state spawning. [ 1048.059962] env[61868]: DEBUG nova.compute.manager [req-687e3c80-1a1d-4229-ad10-07e27f131b5f req-74ac6fdc-3247-4f3d-9627-8da14089c8bf service nova] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Received event network-changed-32232776-6205-48a9-abcc-b24add5b69a1 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1048.060634] env[61868]: DEBUG nova.compute.manager [req-687e3c80-1a1d-4229-ad10-07e27f131b5f req-74ac6fdc-3247-4f3d-9627-8da14089c8bf service nova] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Refreshing instance network info cache due to event network-changed-32232776-6205-48a9-abcc-b24add5b69a1. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1048.060634] env[61868]: DEBUG oslo_concurrency.lockutils [req-687e3c80-1a1d-4229-ad10-07e27f131b5f req-74ac6fdc-3247-4f3d-9627-8da14089c8bf service nova] Acquiring lock "refresh_cache-3a94eb16-cce3-4765-bc61-b6776bdd5b63" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.060789] env[61868]: DEBUG oslo_concurrency.lockutils [req-687e3c80-1a1d-4229-ad10-07e27f131b5f req-74ac6fdc-3247-4f3d-9627-8da14089c8bf service nova] Acquired lock "refresh_cache-3a94eb16-cce3-4765-bc61-b6776bdd5b63" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.060855] env[61868]: DEBUG nova.network.neutron [req-687e3c80-1a1d-4229-ad10-07e27f131b5f req-74ac6fdc-3247-4f3d-9627-8da14089c8bf service nova] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Refreshing network info cache for port 32232776-6205-48a9-abcc-b24add5b69a1 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1048.219504] env[61868]: DEBUG oslo_vmware.api [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52611103-60f8-e6ce-5401-80d082752a7a, 'name': SearchDatastore_Task, 'duration_secs': 0.010391} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.219798] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.220030] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1/7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1048.220303] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cfe60987-8afc-48d5-90e1-87cefd516060 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.227200] env[61868]: DEBUG oslo_vmware.api [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 1048.227200] env[61868]: value = "task-1315939" [ 1048.227200] env[61868]: _type = "Task" [ 1048.227200] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.235315] env[61868]: DEBUG oslo_vmware.api [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315939, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.385947] env[61868]: DEBUG oslo_concurrency.lockutils [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "refresh_cache-3a94eb16-cce3-4765-bc61-b6776bdd5b63" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.405373] env[61868]: DEBUG oslo_vmware.api [None req-fd7d4ac0-445f-48d4-98fe-3fdc7b61e43c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315938, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.560109] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3106b3fa-ff7a-469a-91f0-d93cf49ed972 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.572457] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b74aeac-29a1-4f9a-93f1-6e536da754b1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.602997] env[61868]: DEBUG nova.network.neutron [req-687e3c80-1a1d-4229-ad10-07e27f131b5f req-74ac6fdc-3247-4f3d-9627-8da14089c8bf service nova] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1048.605572] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ed87b15-16ff-4a22-bb7e-dae749a752fc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.615417] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c102196-cadc-4a07-8886-fed0b1cc9813 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.634658] env[61868]: DEBUG nova.compute.provider_tree [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1048.697833] env[61868]: DEBUG nova.network.neutron [req-687e3c80-1a1d-4229-ad10-07e27f131b5f req-74ac6fdc-3247-4f3d-9627-8da14089c8bf service nova] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1048.737780] env[61868]: DEBUG oslo_vmware.api [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315939, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.484074} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.738277] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1/7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1048.738562] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1048.738832] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-031cb1b0-84f2-4835-bb12-793a4a749b53 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.744863] env[61868]: DEBUG oslo_vmware.api [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 1048.744863] env[61868]: value = "task-1315940" [ 1048.744863] env[61868]: _type = "Task" [ 1048.744863] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.752615] env[61868]: DEBUG oslo_vmware.api [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315940, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.902973] env[61868]: DEBUG oslo_vmware.api [None req-fd7d4ac0-445f-48d4-98fe-3fdc7b61e43c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315938, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.138247] env[61868]: DEBUG nova.scheduler.client.report [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1049.200507] env[61868]: DEBUG oslo_concurrency.lockutils [req-687e3c80-1a1d-4229-ad10-07e27f131b5f req-74ac6fdc-3247-4f3d-9627-8da14089c8bf service nova] Releasing lock "refresh_cache-3a94eb16-cce3-4765-bc61-b6776bdd5b63" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.200901] env[61868]: DEBUG oslo_concurrency.lockutils [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquired lock "refresh_cache-3a94eb16-cce3-4765-bc61-b6776bdd5b63" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1049.201081] env[61868]: DEBUG nova.network.neutron [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1049.254151] env[61868]: DEBUG oslo_vmware.api [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315940, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064361} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.254481] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1049.255282] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a020dd08-d539-4fc7-b58c-2a0b7265efba {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.278926] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1/7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1049.279590] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-77d88985-8d19-4b09-b6b4-7baeb67fb2bb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.298517] env[61868]: DEBUG oslo_vmware.api [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 1049.298517] env[61868]: value = "task-1315941" [ 1049.298517] env[61868]: _type = "Task" [ 1049.298517] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.309967] env[61868]: DEBUG oslo_vmware.api [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315941, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.405109] env[61868]: DEBUG oslo_vmware.api [None req-fd7d4ac0-445f-48d4-98fe-3fdc7b61e43c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315938, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.644451] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.178s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.644451] env[61868]: DEBUG nova.compute.manager [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1049.731253] env[61868]: DEBUG nova.network.neutron [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1049.808640] env[61868]: DEBUG oslo_vmware.api [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315941, 'name': ReconfigVM_Task, 'duration_secs': 0.295122} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.808968] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Reconfigured VM instance instance-00000063 to attach disk [datastore2] 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1/7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1049.809605] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a04fadd6-2b83-4d2c-a5b7-06a2e3d52070 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.815781] env[61868]: DEBUG oslo_vmware.api [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 1049.815781] env[61868]: value = "task-1315942" [ 1049.815781] env[61868]: _type = "Task" [ 1049.815781] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.825285] env[61868]: DEBUG oslo_vmware.api [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315942, 'name': Rename_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.852774] env[61868]: DEBUG nova.network.neutron [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Updating instance_info_cache with network_info: [{"id": "32232776-6205-48a9-abcc-b24add5b69a1", "address": "fa:16:3e:5d:60:19", "network": {"id": "77f33b52-e4eb-4b72-85d4-5991a2b5c36d", "bridge": "br-int", "label": "tempest-ServersTestJSON-748635995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b520c3ac58074e8d9b0bfafb817244a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32232776-62", "ovs_interfaceid": "32232776-6205-48a9-abcc-b24add5b69a1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.905624] env[61868]: DEBUG oslo_vmware.api [None req-fd7d4ac0-445f-48d4-98fe-3fdc7b61e43c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315938, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.148839] env[61868]: DEBUG nova.compute.utils [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1050.150352] env[61868]: DEBUG nova.compute.manager [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1050.150561] env[61868]: DEBUG nova.network.neutron [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1050.175841] env[61868]: DEBUG oslo_concurrency.lockutils [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "273a38e7-8eb0-46fc-b9e2-31a9a7fc34db" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.176110] env[61868]: DEBUG oslo_concurrency.lockutils [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "273a38e7-8eb0-46fc-b9e2-31a9a7fc34db" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.188610] env[61868]: DEBUG nova.policy [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1c831b88a7f242469fb3e43955944ba2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '397c2fbb58494513b21b0c638671a5cc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 1050.325999] env[61868]: DEBUG oslo_vmware.api [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315942, 'name': Rename_Task, 'duration_secs': 0.129048} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.326314] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1050.326528] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-40f43fb4-c32c-43bb-927b-41da1d56e432 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.332452] env[61868]: DEBUG oslo_vmware.api [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 1050.332452] env[61868]: value = "task-1315943" [ 1050.332452] env[61868]: _type = "Task" [ 1050.332452] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.339437] env[61868]: DEBUG oslo_vmware.api [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315943, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.355015] env[61868]: DEBUG oslo_concurrency.lockutils [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Releasing lock "refresh_cache-3a94eb16-cce3-4765-bc61-b6776bdd5b63" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1050.355323] env[61868]: DEBUG nova.compute.manager [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Instance network_info: |[{"id": "32232776-6205-48a9-abcc-b24add5b69a1", "address": "fa:16:3e:5d:60:19", "network": {"id": "77f33b52-e4eb-4b72-85d4-5991a2b5c36d", "bridge": "br-int", "label": "tempest-ServersTestJSON-748635995-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b520c3ac58074e8d9b0bfafb817244a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32232776-62", "ovs_interfaceid": "32232776-6205-48a9-abcc-b24add5b69a1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1050.355695] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5d:60:19', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56136ef6-99d7-4562-9a9f-d66fec951c5c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '32232776-6205-48a9-abcc-b24add5b69a1', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1050.363362] env[61868]: DEBUG oslo.service.loopingcall [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1050.363549] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1050.363732] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c757ff02-ed04-4c79-aedf-2a02e17eb243 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.382336] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1050.382336] env[61868]: value = "task-1315944" [ 1050.382336] env[61868]: _type = "Task" [ 1050.382336] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.389723] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315944, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.404418] env[61868]: DEBUG oslo_vmware.api [None req-fd7d4ac0-445f-48d4-98fe-3fdc7b61e43c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315938, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.430028] env[61868]: DEBUG nova.network.neutron [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Successfully created port: 0d4ba939-0509-4e44-885e-2684970089d3 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1050.654116] env[61868]: DEBUG nova.compute.manager [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1050.680243] env[61868]: DEBUG nova.compute.manager [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1050.843076] env[61868]: DEBUG oslo_vmware.api [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315943, 'name': PowerOnVM_Task, 'duration_secs': 0.458691} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.843396] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1050.843798] env[61868]: INFO nova.compute.manager [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Took 7.18 seconds to spawn the instance on the hypervisor. [ 1050.843798] env[61868]: DEBUG nova.compute.manager [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1050.844542] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff8190a1-4e09-42ce-8ed9-735b43fb064e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.891369] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315944, 'name': CreateVM_Task, 'duration_secs': 0.304962} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.891541] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1050.892221] env[61868]: DEBUG oslo_concurrency.lockutils [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1050.892395] env[61868]: DEBUG oslo_concurrency.lockutils [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.892699] env[61868]: DEBUG oslo_concurrency.lockutils [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1050.893009] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d9b40d08-8b88-4cb8-9874-df652488df54 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.899702] env[61868]: DEBUG oslo_vmware.api [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 1050.899702] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5207af49-07da-9a4f-2ede-a03cefc90146" [ 1050.899702] env[61868]: _type = "Task" [ 1050.899702] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.907261] env[61868]: DEBUG oslo_vmware.api [None req-fd7d4ac0-445f-48d4-98fe-3fdc7b61e43c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315938, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.911996] env[61868]: DEBUG oslo_vmware.api [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5207af49-07da-9a4f-2ede-a03cefc90146, 'name': SearchDatastore_Task, 'duration_secs': 0.009625} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.912271] env[61868]: DEBUG oslo_concurrency.lockutils [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1050.912495] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1050.912719] env[61868]: DEBUG oslo_concurrency.lockutils [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1050.912866] env[61868]: DEBUG oslo_concurrency.lockutils [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.913139] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1050.913373] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ef74f6e2-d7f5-4071-ad68-c7b7eae68fc5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.920028] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1050.920214] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1050.920868] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0619219c-6db1-4356-a88a-237f3d609201 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.925605] env[61868]: DEBUG oslo_vmware.api [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 1050.925605] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5293fa3c-566a-dc3e-cea6-938c903e978e" [ 1050.925605] env[61868]: _type = "Task" [ 1050.925605] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.932598] env[61868]: DEBUG oslo_vmware.api [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5293fa3c-566a-dc3e-cea6-938c903e978e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.204046] env[61868]: DEBUG oslo_concurrency.lockutils [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.204359] env[61868]: DEBUG oslo_concurrency.lockutils [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.205871] env[61868]: INFO nova.compute.claims [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1051.362207] env[61868]: INFO nova.compute.manager [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Took 14.35 seconds to build instance. [ 1051.404934] env[61868]: DEBUG oslo_vmware.api [None req-fd7d4ac0-445f-48d4-98fe-3fdc7b61e43c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315938, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.435449] env[61868]: DEBUG oslo_vmware.api [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5293fa3c-566a-dc3e-cea6-938c903e978e, 'name': SearchDatastore_Task, 'duration_secs': 0.010103} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.436220] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-422d3a47-1356-4de1-859f-9162c0d634f3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.441524] env[61868]: DEBUG oslo_vmware.api [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 1051.441524] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]527b98bd-f38e-31d8-ffa5-4a65510084e2" [ 1051.441524] env[61868]: _type = "Task" [ 1051.441524] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.449160] env[61868]: DEBUG oslo_vmware.api [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]527b98bd-f38e-31d8-ffa5-4a65510084e2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.663947] env[61868]: DEBUG nova.compute.manager [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1051.690953] env[61868]: DEBUG nova.virt.hardware [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1051.693272] env[61868]: DEBUG nova.virt.hardware [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1051.693272] env[61868]: DEBUG nova.virt.hardware [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1051.693272] env[61868]: DEBUG nova.virt.hardware [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1051.693272] env[61868]: DEBUG nova.virt.hardware [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1051.693272] env[61868]: DEBUG nova.virt.hardware [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1051.693272] env[61868]: DEBUG nova.virt.hardware [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1051.693272] env[61868]: DEBUG nova.virt.hardware [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1051.693272] env[61868]: DEBUG nova.virt.hardware [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1051.693272] env[61868]: DEBUG nova.virt.hardware [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1051.693272] env[61868]: DEBUG nova.virt.hardware [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1051.693681] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8339697c-7bb1-41b4-b208-d1d301e9e420 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.701575] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97f79eac-e27a-4d7b-8cd2-ed1e104f6c5e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.864406] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a881ce95-f493-4853-b600-27be2053fa5b tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.856s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.906378] env[61868]: DEBUG oslo_vmware.api [None req-fd7d4ac0-445f-48d4-98fe-3fdc7b61e43c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315938, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.951749] env[61868]: DEBUG oslo_vmware.api [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]527b98bd-f38e-31d8-ffa5-4a65510084e2, 'name': SearchDatastore_Task, 'duration_secs': 0.010419} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.952076] env[61868]: DEBUG oslo_concurrency.lockutils [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1051.952384] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 3a94eb16-cce3-4765-bc61-b6776bdd5b63/3a94eb16-cce3-4765-bc61-b6776bdd5b63.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1051.952665] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-305d5c77-4256-4f8a-9d1a-1403d5a7ada1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.959133] env[61868]: DEBUG oslo_vmware.api [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 1051.959133] env[61868]: value = "task-1315945" [ 1051.959133] env[61868]: _type = "Task" [ 1051.959133] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.966595] env[61868]: DEBUG oslo_vmware.api [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315945, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.053380] env[61868]: DEBUG nova.compute.manager [req-8830023f-ad2b-4f63-84fb-8c96307b3866 req-37afcb7e-a114-4450-8b85-64a2cd773655 service nova] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Received event network-vif-plugged-0d4ba939-0509-4e44-885e-2684970089d3 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1052.053602] env[61868]: DEBUG oslo_concurrency.lockutils [req-8830023f-ad2b-4f63-84fb-8c96307b3866 req-37afcb7e-a114-4450-8b85-64a2cd773655 service nova] Acquiring lock "1cb04c98-f725-4b1f-933e-ded580d340fb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.053816] env[61868]: DEBUG oslo_concurrency.lockutils [req-8830023f-ad2b-4f63-84fb-8c96307b3866 req-37afcb7e-a114-4450-8b85-64a2cd773655 service nova] Lock "1cb04c98-f725-4b1f-933e-ded580d340fb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.053992] env[61868]: DEBUG oslo_concurrency.lockutils [req-8830023f-ad2b-4f63-84fb-8c96307b3866 req-37afcb7e-a114-4450-8b85-64a2cd773655 service nova] Lock "1cb04c98-f725-4b1f-933e-ded580d340fb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.054217] env[61868]: DEBUG nova.compute.manager [req-8830023f-ad2b-4f63-84fb-8c96307b3866 req-37afcb7e-a114-4450-8b85-64a2cd773655 service nova] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] No waiting events found dispatching network-vif-plugged-0d4ba939-0509-4e44-885e-2684970089d3 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1052.054394] env[61868]: WARNING nova.compute.manager [req-8830023f-ad2b-4f63-84fb-8c96307b3866 req-37afcb7e-a114-4450-8b85-64a2cd773655 service nova] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Received unexpected event network-vif-plugged-0d4ba939-0509-4e44-885e-2684970089d3 for instance with vm_state building and task_state spawning. [ 1052.167956] env[61868]: DEBUG nova.network.neutron [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Successfully updated port: 0d4ba939-0509-4e44-885e-2684970089d3 {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1052.320558] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02b19bf0-3d4d-4690-94f0-3803eeefb9ea {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.328459] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81817ece-1d73-48d2-a6ed-328c0e3397d0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.359517] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5439ca6c-3d5d-40a0-a839-e40b72e3833f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.367442] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64a31ede-5987-4138-97f3-838d17063185 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.382877] env[61868]: DEBUG nova.compute.provider_tree [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1052.406602] env[61868]: DEBUG oslo_vmware.api [None req-fd7d4ac0-445f-48d4-98fe-3fdc7b61e43c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315938, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.468637] env[61868]: DEBUG oslo_vmware.api [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315945, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.44085} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.468889] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 3a94eb16-cce3-4765-bc61-b6776bdd5b63/3a94eb16-cce3-4765-bc61-b6776bdd5b63.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1052.469140] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1052.469382] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1ae14511-22b1-4bb6-9d66-8dc6359d8dbf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.475156] env[61868]: DEBUG oslo_vmware.api [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 1052.475156] env[61868]: value = "task-1315946" [ 1052.475156] env[61868]: _type = "Task" [ 1052.475156] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.483574] env[61868]: DEBUG oslo_vmware.api [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315946, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.671514] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Acquiring lock "refresh_cache-1cb04c98-f725-4b1f-933e-ded580d340fb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1052.671779] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Acquired lock "refresh_cache-1cb04c98-f725-4b1f-933e-ded580d340fb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1052.671779] env[61868]: DEBUG nova.network.neutron [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1052.886110] env[61868]: DEBUG nova.scheduler.client.report [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1052.907885] env[61868]: DEBUG oslo_vmware.api [None req-fd7d4ac0-445f-48d4-98fe-3fdc7b61e43c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315938, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.984430] env[61868]: DEBUG oslo_vmware.api [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315946, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06115} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.984638] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1052.985424] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-672c8f9c-b62b-48d0-9d72-a0769e29f9bc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.006993] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] 3a94eb16-cce3-4765-bc61-b6776bdd5b63/3a94eb16-cce3-4765-bc61-b6776bdd5b63.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1053.007347] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-47f6a1f8-318c-4b20-8f86-cceac7de87cd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.021959] env[61868]: DEBUG nova.compute.manager [None req-9010f905-ecb4-4531-ab1a-1061c24fb7bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1053.022715] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec7af17d-5d82-45fd-a555-f6e3accb77e8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.032893] env[61868]: DEBUG oslo_vmware.api [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 1053.032893] env[61868]: value = "task-1315947" [ 1053.032893] env[61868]: _type = "Task" [ 1053.032893] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.041886] env[61868]: DEBUG oslo_vmware.api [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315947, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.204494] env[61868]: DEBUG nova.network.neutron [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1053.346011] env[61868]: DEBUG nova.network.neutron [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Updating instance_info_cache with network_info: [{"id": "0d4ba939-0509-4e44-885e-2684970089d3", "address": "fa:16:3e:32:33:5a", "network": {"id": "01b566bd-7401-4a8e-94dd-82ba93836282", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1185865401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "397c2fbb58494513b21b0c638671a5cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9297313e-7c50-4873-93d3-67284929163a", "external-id": "nsx-vlan-transportzone-620", "segmentation_id": 620, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0d4ba939-05", "ovs_interfaceid": "0d4ba939-0509-4e44-885e-2684970089d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.391302] env[61868]: DEBUG oslo_concurrency.lockutils [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.187s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.391828] env[61868]: DEBUG nova.compute.manager [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1053.406788] env[61868]: DEBUG oslo_vmware.api [None req-fd7d4ac0-445f-48d4-98fe-3fdc7b61e43c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315938, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.535192] env[61868]: INFO nova.compute.manager [None req-9010f905-ecb4-4531-ab1a-1061c24fb7bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] instance snapshotting [ 1053.541439] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2836c3e9-3376-4327-92de-ee59a3b4933a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.550409] env[61868]: DEBUG oslo_vmware.api [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315947, 'name': ReconfigVM_Task, 'duration_secs': 0.304162} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.571855] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Reconfigured VM instance instance-00000064 to attach disk [datastore2] 3a94eb16-cce3-4765-bc61-b6776bdd5b63/3a94eb16-cce3-4765-bc61-b6776bdd5b63.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1053.572869] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-94ec3191-b69b-4982-905c-f5586ca7a81a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.574950] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3835e7b-24de-4336-a220-386248e38296 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.584848] env[61868]: DEBUG oslo_vmware.api [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 1053.584848] env[61868]: value = "task-1315948" [ 1053.584848] env[61868]: _type = "Task" [ 1053.584848] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.592585] env[61868]: DEBUG oslo_vmware.api [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315948, 'name': Rename_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.849026] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Releasing lock "refresh_cache-1cb04c98-f725-4b1f-933e-ded580d340fb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1053.849026] env[61868]: DEBUG nova.compute.manager [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Instance network_info: |[{"id": "0d4ba939-0509-4e44-885e-2684970089d3", "address": "fa:16:3e:32:33:5a", "network": {"id": "01b566bd-7401-4a8e-94dd-82ba93836282", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1185865401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "397c2fbb58494513b21b0c638671a5cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9297313e-7c50-4873-93d3-67284929163a", "external-id": "nsx-vlan-transportzone-620", "segmentation_id": 620, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0d4ba939-05", "ovs_interfaceid": "0d4ba939-0509-4e44-885e-2684970089d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1053.849295] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:32:33:5a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9297313e-7c50-4873-93d3-67284929163a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0d4ba939-0509-4e44-885e-2684970089d3', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1053.856545] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Creating folder: Project (397c2fbb58494513b21b0c638671a5cc). Parent ref: group-v281478. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1053.856836] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-90d3fac0-32f3-46ab-9848-ee4e83c61415 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.867036] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Created folder: Project (397c2fbb58494513b21b0c638671a5cc) in parent group-v281478. [ 1053.867137] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Creating folder: Instances. Parent ref: group-v281615. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1053.867321] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6bed32d5-ccd8-4034-936c-658e7fa8c96a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.876949] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Created folder: Instances in parent group-v281615. [ 1053.877185] env[61868]: DEBUG oslo.service.loopingcall [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1053.877371] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1053.877564] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8e782ae3-cd09-4b83-bb55-7b8df4cb617e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.897157] env[61868]: DEBUG nova.compute.utils [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1053.898413] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1053.898413] env[61868]: value = "task-1315951" [ 1053.898413] env[61868]: _type = "Task" [ 1053.898413] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.898933] env[61868]: DEBUG nova.compute.manager [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1053.898933] env[61868]: DEBUG nova.network.neutron [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1053.912628] env[61868]: DEBUG oslo_vmware.api [None req-fd7d4ac0-445f-48d4-98fe-3fdc7b61e43c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315938, 'name': ReconfigVM_Task, 'duration_secs': 5.955468} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.915778] env[61868]: DEBUG oslo_concurrency.lockutils [None req-fd7d4ac0-445f-48d4-98fe-3fdc7b61e43c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Releasing lock "69a0b94b-2bf4-49ad-8fb0-ea945176e5eb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1053.916009] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-fd7d4ac0-445f-48d4-98fe-3fdc7b61e43c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Reconfigured VM to detach interface {{(pid=61868) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1053.918233] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315951, 'name': CreateVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.947225] env[61868]: DEBUG nova.policy [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0fa71d6f8941441d89709e7e3af1ed9f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c8440ce785c344d793dd513b5f008d65', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 1054.080671] env[61868]: DEBUG nova.compute.manager [req-db97a9ee-45c4-44be-80bd-e3d682047274 req-d2fa7650-135f-43c4-b25a-49bddf3dfb35 service nova] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Received event network-changed-0d4ba939-0509-4e44-885e-2684970089d3 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1054.080951] env[61868]: DEBUG nova.compute.manager [req-db97a9ee-45c4-44be-80bd-e3d682047274 req-d2fa7650-135f-43c4-b25a-49bddf3dfb35 service nova] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Refreshing instance network info cache due to event network-changed-0d4ba939-0509-4e44-885e-2684970089d3. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1054.081319] env[61868]: DEBUG oslo_concurrency.lockutils [req-db97a9ee-45c4-44be-80bd-e3d682047274 req-d2fa7650-135f-43c4-b25a-49bddf3dfb35 service nova] Acquiring lock "refresh_cache-1cb04c98-f725-4b1f-933e-ded580d340fb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.081491] env[61868]: DEBUG oslo_concurrency.lockutils [req-db97a9ee-45c4-44be-80bd-e3d682047274 req-d2fa7650-135f-43c4-b25a-49bddf3dfb35 service nova] Acquired lock "refresh_cache-1cb04c98-f725-4b1f-933e-ded580d340fb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.081668] env[61868]: DEBUG nova.network.neutron [req-db97a9ee-45c4-44be-80bd-e3d682047274 req-d2fa7650-135f-43c4-b25a-49bddf3dfb35 service nova] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Refreshing network info cache for port 0d4ba939-0509-4e44-885e-2684970089d3 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1054.086462] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9010f905-ecb4-4531-ab1a-1061c24fb7bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Creating Snapshot of the VM instance {{(pid=61868) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1054.087050] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-db505235-bf61-472f-b6ac-e1347f674246 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.099171] env[61868]: DEBUG oslo_vmware.api [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315948, 'name': Rename_Task, 'duration_secs': 0.131309} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.100485] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1054.100805] env[61868]: DEBUG oslo_vmware.api [None req-9010f905-ecb4-4531-ab1a-1061c24fb7bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 1054.100805] env[61868]: value = "task-1315952" [ 1054.100805] env[61868]: _type = "Task" [ 1054.100805] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.101038] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f5bdb9ac-514f-41d3-b4d6-15b1868e21d7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.111547] env[61868]: DEBUG oslo_vmware.api [None req-9010f905-ecb4-4531-ab1a-1061c24fb7bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315952, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.113574] env[61868]: DEBUG oslo_vmware.api [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 1054.113574] env[61868]: value = "task-1315953" [ 1054.113574] env[61868]: _type = "Task" [ 1054.113574] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.122624] env[61868]: DEBUG oslo_vmware.api [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315953, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.257525] env[61868]: DEBUG nova.network.neutron [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Successfully created port: a34b4784-533a-42f9-b5c9-ea8332e8eecb {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1054.401669] env[61868]: DEBUG nova.compute.manager [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1054.427239] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315951, 'name': CreateVM_Task, 'duration_secs': 0.474292} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.427374] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1054.428192] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.428414] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.428896] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1054.430982] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-94589257-68d9-4915-a46c-92c1be550d15 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.436761] env[61868]: DEBUG oslo_vmware.api [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Waiting for the task: (returnval){ [ 1054.436761] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]529964a6-6b72-f868-7216-1f11c2ec0571" [ 1054.436761] env[61868]: _type = "Task" [ 1054.436761] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.449946] env[61868]: DEBUG oslo_vmware.api [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]529964a6-6b72-f868-7216-1f11c2ec0571, 'name': SearchDatastore_Task, 'duration_secs': 0.009337} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.450272] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1054.450737] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1054.451139] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.451276] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.451472] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1054.451740] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6aec841c-4944-4950-ab04-2532ba6ca6f2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.459813] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1054.459994] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1054.460836] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82adba38-fdc9-4909-ab5e-0cf508136369 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.466555] env[61868]: DEBUG oslo_vmware.api [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Waiting for the task: (returnval){ [ 1054.466555] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52428bd3-b6c5-6211-f051-76c4525ad675" [ 1054.466555] env[61868]: _type = "Task" [ 1054.466555] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.474088] env[61868]: DEBUG oslo_vmware.api [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52428bd3-b6c5-6211-f051-76c4525ad675, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.612283] env[61868]: DEBUG oslo_vmware.api [None req-9010f905-ecb4-4531-ab1a-1061c24fb7bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315952, 'name': CreateSnapshot_Task, 'duration_secs': 0.511665} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.612434] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9010f905-ecb4-4531-ab1a-1061c24fb7bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Created Snapshot of the VM instance {{(pid=61868) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1054.613135] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6003130-d703-43fa-9437-1756ef949af1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.629531] env[61868]: DEBUG oslo_vmware.api [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315953, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.780296] env[61868]: DEBUG nova.network.neutron [req-db97a9ee-45c4-44be-80bd-e3d682047274 req-d2fa7650-135f-43c4-b25a-49bddf3dfb35 service nova] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Updated VIF entry in instance network info cache for port 0d4ba939-0509-4e44-885e-2684970089d3. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1054.780708] env[61868]: DEBUG nova.network.neutron [req-db97a9ee-45c4-44be-80bd-e3d682047274 req-d2fa7650-135f-43c4-b25a-49bddf3dfb35 service nova] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Updating instance_info_cache with network_info: [{"id": "0d4ba939-0509-4e44-885e-2684970089d3", "address": "fa:16:3e:32:33:5a", "network": {"id": "01b566bd-7401-4a8e-94dd-82ba93836282", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1185865401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "397c2fbb58494513b21b0c638671a5cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9297313e-7c50-4873-93d3-67284929163a", "external-id": "nsx-vlan-transportzone-620", "segmentation_id": 620, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0d4ba939-05", "ovs_interfaceid": "0d4ba939-0509-4e44-885e-2684970089d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.980349] env[61868]: DEBUG oslo_vmware.api [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52428bd3-b6c5-6211-f051-76c4525ad675, 'name': SearchDatastore_Task, 'duration_secs': 0.00924} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.981586] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7426a9e1-0191-43a6-a5ad-ac5885359b65 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.987830] env[61868]: DEBUG oslo_vmware.api [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Waiting for the task: (returnval){ [ 1054.987830] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e014e6-14e4-f5e6-4281-1fdf172a3d57" [ 1054.987830] env[61868]: _type = "Task" [ 1054.987830] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.995835] env[61868]: DEBUG oslo_vmware.api [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e014e6-14e4-f5e6-4281-1fdf172a3d57, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.125895] env[61868]: DEBUG oslo_vmware.api [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315953, 'name': PowerOnVM_Task, 'duration_secs': 0.890348} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.126211] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1055.126504] env[61868]: INFO nova.compute.manager [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Took 7.56 seconds to spawn the instance on the hypervisor. [ 1055.126693] env[61868]: DEBUG nova.compute.manager [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1055.135098] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9010f905-ecb4-4531-ab1a-1061c24fb7bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Creating linked-clone VM from snapshot {{(pid=61868) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1055.135854] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-643f69e6-d9c2-4c66-aa99-1e7e651f01f6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.138581] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-b92d03b9-e535-4ed0-8a9b-51e68c61e9f6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.148856] env[61868]: DEBUG oslo_vmware.api [None req-9010f905-ecb4-4531-ab1a-1061c24fb7bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 1055.148856] env[61868]: value = "task-1315954" [ 1055.148856] env[61868]: _type = "Task" [ 1055.148856] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.156908] env[61868]: DEBUG oslo_vmware.api [None req-9010f905-ecb4-4531-ab1a-1061c24fb7bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315954, 'name': CloneVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.284209] env[61868]: DEBUG oslo_concurrency.lockutils [req-db97a9ee-45c4-44be-80bd-e3d682047274 req-d2fa7650-135f-43c4-b25a-49bddf3dfb35 service nova] Releasing lock "refresh_cache-1cb04c98-f725-4b1f-933e-ded580d340fb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1055.421418] env[61868]: DEBUG nova.compute.manager [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1055.464780] env[61868]: DEBUG nova.virt.hardware [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1055.465053] env[61868]: DEBUG nova.virt.hardware [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1055.465225] env[61868]: DEBUG nova.virt.hardware [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1055.465423] env[61868]: DEBUG nova.virt.hardware [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1055.465579] env[61868]: DEBUG nova.virt.hardware [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1055.465743] env[61868]: DEBUG nova.virt.hardware [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1055.466238] env[61868]: DEBUG nova.virt.hardware [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1055.466238] env[61868]: DEBUG nova.virt.hardware [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1055.466351] env[61868]: DEBUG nova.virt.hardware [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1055.466510] env[61868]: DEBUG nova.virt.hardware [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1055.466698] env[61868]: DEBUG nova.virt.hardware [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1055.467646] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87337d41-fb71-4d95-ba30-48271df52431 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.475749] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3496c26-32e5-4c5b-9606-9ad66b55c01c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.500420] env[61868]: DEBUG oslo_vmware.api [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e014e6-14e4-f5e6-4281-1fdf172a3d57, 'name': SearchDatastore_Task, 'duration_secs': 0.030371} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.500718] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1055.501030] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 1cb04c98-f725-4b1f-933e-ded580d340fb/1cb04c98-f725-4b1f-933e-ded580d340fb.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1055.501301] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f98cc270-4eec-4676-89c4-314091034db1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.508295] env[61868]: DEBUG oslo_vmware.api [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Waiting for the task: (returnval){ [ 1055.508295] env[61868]: value = "task-1315955" [ 1055.508295] env[61868]: _type = "Task" [ 1055.508295] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.517262] env[61868]: DEBUG oslo_vmware.api [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Task: {'id': task-1315955, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.578094] env[61868]: DEBUG oslo_concurrency.lockutils [None req-fd7d4ac0-445f-48d4-98fe-3fdc7b61e43c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "refresh_cache-69a0b94b-2bf4-49ad-8fb0-ea945176e5eb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1055.578387] env[61868]: DEBUG oslo_concurrency.lockutils [None req-fd7d4ac0-445f-48d4-98fe-3fdc7b61e43c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquired lock "refresh_cache-69a0b94b-2bf4-49ad-8fb0-ea945176e5eb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1055.578606] env[61868]: DEBUG nova.network.neutron [None req-fd7d4ac0-445f-48d4-98fe-3fdc7b61e43c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1055.661929] env[61868]: INFO nova.compute.manager [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Took 15.23 seconds to build instance. [ 1055.669160] env[61868]: DEBUG oslo_vmware.api [None req-9010f905-ecb4-4531-ab1a-1061c24fb7bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315954, 'name': CloneVM_Task} progress is 94%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.760683] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bd363ea7-bc29-46fa-adb7-999d56da820c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "69a0b94b-2bf4-49ad-8fb0-ea945176e5eb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.761062] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bd363ea7-bc29-46fa-adb7-999d56da820c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "69a0b94b-2bf4-49ad-8fb0-ea945176e5eb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.761294] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bd363ea7-bc29-46fa-adb7-999d56da820c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "69a0b94b-2bf4-49ad-8fb0-ea945176e5eb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.761488] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bd363ea7-bc29-46fa-adb7-999d56da820c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "69a0b94b-2bf4-49ad-8fb0-ea945176e5eb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.761680] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bd363ea7-bc29-46fa-adb7-999d56da820c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "69a0b94b-2bf4-49ad-8fb0-ea945176e5eb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.765271] env[61868]: INFO nova.compute.manager [None req-bd363ea7-bc29-46fa-adb7-999d56da820c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Terminating instance [ 1055.769419] env[61868]: DEBUG nova.compute.manager [None req-bd363ea7-bc29-46fa-adb7-999d56da820c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1055.769798] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-bd363ea7-bc29-46fa-adb7-999d56da820c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1055.770621] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58c62ce2-04c8-4de1-8047-65b5edb5e21f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.780614] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd363ea7-bc29-46fa-adb7-999d56da820c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1055.780991] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5dd92ea6-5f78-497b-90ea-ac9e84800352 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.785098] env[61868]: DEBUG nova.network.neutron [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Successfully updated port: a34b4784-533a-42f9-b5c9-ea8332e8eecb {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1055.791894] env[61868]: DEBUG oslo_vmware.api [None req-bd363ea7-bc29-46fa-adb7-999d56da820c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for the task: (returnval){ [ 1055.791894] env[61868]: value = "task-1315956" [ 1055.791894] env[61868]: _type = "Task" [ 1055.791894] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.804426] env[61868]: DEBUG oslo_vmware.api [None req-bd363ea7-bc29-46fa-adb7-999d56da820c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315956, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.019825] env[61868]: DEBUG oslo_vmware.api [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Task: {'id': task-1315955, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.163872] env[61868]: DEBUG oslo_vmware.api [None req-9010f905-ecb4-4531-ab1a-1061c24fb7bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315954, 'name': CloneVM_Task} progress is 94%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.164322] env[61868]: DEBUG oslo_concurrency.lockutils [None req-57aab834-4d1a-4cc8-9733-bedab0d99c6e tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "3a94eb16-cce3-4765-bc61-b6776bdd5b63" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.740s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.166798] env[61868]: DEBUG nova.compute.manager [req-c54feadb-f4bb-46a3-92d5-565e4d383a1d req-61e118d9-75f5-4aff-8132-be6857d006ac service nova] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Received event network-vif-plugged-a34b4784-533a-42f9-b5c9-ea8332e8eecb {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1056.166993] env[61868]: DEBUG oslo_concurrency.lockutils [req-c54feadb-f4bb-46a3-92d5-565e4d383a1d req-61e118d9-75f5-4aff-8132-be6857d006ac service nova] Acquiring lock "273a38e7-8eb0-46fc-b9e2-31a9a7fc34db-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.167220] env[61868]: DEBUG oslo_concurrency.lockutils [req-c54feadb-f4bb-46a3-92d5-565e4d383a1d req-61e118d9-75f5-4aff-8132-be6857d006ac service nova] Lock "273a38e7-8eb0-46fc-b9e2-31a9a7fc34db-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.167380] env[61868]: DEBUG oslo_concurrency.lockutils [req-c54feadb-f4bb-46a3-92d5-565e4d383a1d req-61e118d9-75f5-4aff-8132-be6857d006ac service nova] Lock "273a38e7-8eb0-46fc-b9e2-31a9a7fc34db-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.167545] env[61868]: DEBUG nova.compute.manager [req-c54feadb-f4bb-46a3-92d5-565e4d383a1d req-61e118d9-75f5-4aff-8132-be6857d006ac service nova] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] No waiting events found dispatching network-vif-plugged-a34b4784-533a-42f9-b5c9-ea8332e8eecb {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1056.167788] env[61868]: WARNING nova.compute.manager [req-c54feadb-f4bb-46a3-92d5-565e4d383a1d req-61e118d9-75f5-4aff-8132-be6857d006ac service nova] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Received unexpected event network-vif-plugged-a34b4784-533a-42f9-b5c9-ea8332e8eecb for instance with vm_state building and task_state spawning. [ 1056.167892] env[61868]: DEBUG nova.compute.manager [req-c54feadb-f4bb-46a3-92d5-565e4d383a1d req-61e118d9-75f5-4aff-8132-be6857d006ac service nova] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Received event network-changed-a34b4784-533a-42f9-b5c9-ea8332e8eecb {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1056.168018] env[61868]: DEBUG nova.compute.manager [req-c54feadb-f4bb-46a3-92d5-565e4d383a1d req-61e118d9-75f5-4aff-8132-be6857d006ac service nova] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Refreshing instance network info cache due to event network-changed-a34b4784-533a-42f9-b5c9-ea8332e8eecb. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1056.168208] env[61868]: DEBUG oslo_concurrency.lockutils [req-c54feadb-f4bb-46a3-92d5-565e4d383a1d req-61e118d9-75f5-4aff-8132-be6857d006ac service nova] Acquiring lock "refresh_cache-273a38e7-8eb0-46fc-b9e2-31a9a7fc34db" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1056.168344] env[61868]: DEBUG oslo_concurrency.lockutils [req-c54feadb-f4bb-46a3-92d5-565e4d383a1d req-61e118d9-75f5-4aff-8132-be6857d006ac service nova] Acquired lock "refresh_cache-273a38e7-8eb0-46fc-b9e2-31a9a7fc34db" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.168499] env[61868]: DEBUG nova.network.neutron [req-c54feadb-f4bb-46a3-92d5-565e4d383a1d req-61e118d9-75f5-4aff-8132-be6857d006ac service nova] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Refreshing network info cache for port a34b4784-533a-42f9-b5c9-ea8332e8eecb {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1056.274419] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6ff3b3b2-959e-43b8-91b8-a6664466dcd8 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "3a94eb16-cce3-4765-bc61-b6776bdd5b63" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.275047] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6ff3b3b2-959e-43b8-91b8-a6664466dcd8 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "3a94eb16-cce3-4765-bc61-b6776bdd5b63" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.275317] env[61868]: DEBUG nova.compute.manager [None req-6ff3b3b2-959e-43b8-91b8-a6664466dcd8 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1056.276185] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8ba4283-e2f0-4cb7-bc4f-1d591bcf7721 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.282750] env[61868]: DEBUG nova.compute.manager [None req-6ff3b3b2-959e-43b8-91b8-a6664466dcd8 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61868) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1056.283380] env[61868]: DEBUG nova.objects.instance [None req-6ff3b3b2-959e-43b8-91b8-a6664466dcd8 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lazy-loading 'flavor' on Instance uuid 3a94eb16-cce3-4765-bc61-b6776bdd5b63 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1056.287826] env[61868]: DEBUG oslo_concurrency.lockutils [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "refresh_cache-273a38e7-8eb0-46fc-b9e2-31a9a7fc34db" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1056.311641] env[61868]: DEBUG oslo_vmware.api [None req-bd363ea7-bc29-46fa-adb7-999d56da820c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315956, 'name': PowerOffVM_Task, 'duration_secs': 0.216939} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.312721] env[61868]: INFO nova.network.neutron [None req-fd7d4ac0-445f-48d4-98fe-3fdc7b61e43c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Port 0bbbc97a-2c41-4d5d-a093-fca57a1d543b from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1056.313308] env[61868]: DEBUG nova.network.neutron [None req-fd7d4ac0-445f-48d4-98fe-3fdc7b61e43c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Updating instance_info_cache with network_info: [{"id": "f8f4c521-03cd-4e54-99f1-761d9a515aa3", "address": "fa:16:3e:3c:dd:65", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8f4c521-03", "ovs_interfaceid": "f8f4c521-03cd-4e54-99f1-761d9a515aa3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1056.314413] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd363ea7-bc29-46fa-adb7-999d56da820c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1056.314491] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-bd363ea7-bc29-46fa-adb7-999d56da820c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1056.315011] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2c68fd61-fd9e-4e39-a1a2-8829a04554a3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.385269] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-bd363ea7-bc29-46fa-adb7-999d56da820c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1056.385487] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-bd363ea7-bc29-46fa-adb7-999d56da820c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Deleting contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1056.385678] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd363ea7-bc29-46fa-adb7-999d56da820c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Deleting the datastore file [datastore2] 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1056.385950] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-49eb971e-dfaf-46d5-9ae9-c2ffbc5d7b56 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.392501] env[61868]: DEBUG oslo_vmware.api [None req-bd363ea7-bc29-46fa-adb7-999d56da820c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for the task: (returnval){ [ 1056.392501] env[61868]: value = "task-1315958" [ 1056.392501] env[61868]: _type = "Task" [ 1056.392501] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.400949] env[61868]: DEBUG oslo_vmware.api [None req-bd363ea7-bc29-46fa-adb7-999d56da820c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315958, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.522060] env[61868]: DEBUG oslo_vmware.api [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Task: {'id': task-1315955, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.600738} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.522060] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 1cb04c98-f725-4b1f-933e-ded580d340fb/1cb04c98-f725-4b1f-933e-ded580d340fb.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1056.522060] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1056.522060] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d17bbabb-3fcf-44b8-b1f1-cd96ee04a6b9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.528019] env[61868]: DEBUG oslo_vmware.api [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Waiting for the task: (returnval){ [ 1056.528019] env[61868]: value = "task-1315959" [ 1056.528019] env[61868]: _type = "Task" [ 1056.528019] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.535918] env[61868]: DEBUG oslo_vmware.api [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Task: {'id': task-1315959, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.665717] env[61868]: DEBUG oslo_vmware.api [None req-9010f905-ecb4-4531-ab1a-1061c24fb7bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315954, 'name': CloneVM_Task} progress is 100%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.698827] env[61868]: DEBUG nova.network.neutron [req-c54feadb-f4bb-46a3-92d5-565e4d383a1d req-61e118d9-75f5-4aff-8132-be6857d006ac service nova] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1056.773245] env[61868]: DEBUG nova.network.neutron [req-c54feadb-f4bb-46a3-92d5-565e4d383a1d req-61e118d9-75f5-4aff-8132-be6857d006ac service nova] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1056.788722] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ff3b3b2-959e-43b8-91b8-a6664466dcd8 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1056.788987] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e2ff2967-d9ed-4164-ac56-192c3d291657 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.795862] env[61868]: DEBUG oslo_vmware.api [None req-6ff3b3b2-959e-43b8-91b8-a6664466dcd8 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 1056.795862] env[61868]: value = "task-1315960" [ 1056.795862] env[61868]: _type = "Task" [ 1056.795862] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.804061] env[61868]: DEBUG oslo_vmware.api [None req-6ff3b3b2-959e-43b8-91b8-a6664466dcd8 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315960, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.807550] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1056.807828] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1056.807896] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Starting heal instance info cache {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1056.807999] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Rebuilding the list of instances to heal {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1056.816227] env[61868]: DEBUG oslo_concurrency.lockutils [None req-fd7d4ac0-445f-48d4-98fe-3fdc7b61e43c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Releasing lock "refresh_cache-69a0b94b-2bf4-49ad-8fb0-ea945176e5eb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1056.902736] env[61868]: DEBUG oslo_vmware.api [None req-bd363ea7-bc29-46fa-adb7-999d56da820c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315958, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.040136] env[61868]: DEBUG oslo_vmware.api [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Task: {'id': task-1315959, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.177244} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.040440] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1057.041512] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8fc1a04-2e3f-4160-8362-548024a8cb99 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.072653] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] 1cb04c98-f725-4b1f-933e-ded580d340fb/1cb04c98-f725-4b1f-933e-ded580d340fb.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1057.072999] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e1a9ee05-306f-429c-9eec-71d45266491c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.100438] env[61868]: DEBUG oslo_vmware.api [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Waiting for the task: (returnval){ [ 1057.100438] env[61868]: value = "task-1315961" [ 1057.100438] env[61868]: _type = "Task" [ 1057.100438] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.110444] env[61868]: DEBUG oslo_vmware.api [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Task: {'id': task-1315961, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.166233] env[61868]: DEBUG oslo_vmware.api [None req-9010f905-ecb4-4531-ab1a-1061c24fb7bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315954, 'name': CloneVM_Task, 'duration_secs': 1.675395} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.166521] env[61868]: INFO nova.virt.vmwareapi.vmops [None req-9010f905-ecb4-4531-ab1a-1061c24fb7bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Created linked-clone VM from snapshot [ 1057.167336] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b090f13-d011-4741-9493-d22c2f724cdf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.176506] env[61868]: DEBUG nova.virt.vmwareapi.images [None req-9010f905-ecb4-4531-ab1a-1061c24fb7bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Uploading image ee4de7b2-a1b2-473c-b4f8-a9e13795fe3c {{(pid=61868) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1057.191112] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-9010f905-ecb4-4531-ab1a-1061c24fb7bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Destroying the VM {{(pid=61868) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1057.191431] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-3ef87b2e-ca35-4ccc-8fd1-3adf055bd803 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.198599] env[61868]: DEBUG oslo_vmware.api [None req-9010f905-ecb4-4531-ab1a-1061c24fb7bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 1057.198599] env[61868]: value = "task-1315962" [ 1057.198599] env[61868]: _type = "Task" [ 1057.198599] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.208473] env[61868]: DEBUG oslo_vmware.api [None req-9010f905-ecb4-4531-ab1a-1061c24fb7bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315962, 'name': Destroy_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.276346] env[61868]: DEBUG oslo_concurrency.lockutils [req-c54feadb-f4bb-46a3-92d5-565e4d383a1d req-61e118d9-75f5-4aff-8132-be6857d006ac service nova] Releasing lock "refresh_cache-273a38e7-8eb0-46fc-b9e2-31a9a7fc34db" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1057.276733] env[61868]: DEBUG oslo_concurrency.lockutils [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquired lock "refresh_cache-273a38e7-8eb0-46fc-b9e2-31a9a7fc34db" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.276892] env[61868]: DEBUG nova.network.neutron [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1057.305541] env[61868]: DEBUG oslo_vmware.api [None req-6ff3b3b2-959e-43b8-91b8-a6664466dcd8 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315960, 'name': PowerOffVM_Task, 'duration_secs': 0.437346} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.305779] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ff3b3b2-959e-43b8-91b8-a6664466dcd8 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1057.305959] env[61868]: DEBUG nova.compute.manager [None req-6ff3b3b2-959e-43b8-91b8-a6664466dcd8 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1057.306721] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dc5b062-d912-4203-9445-641df1e7f447 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.312086] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Skipping network cache update for instance because it is being deleted. {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 1057.312250] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Skipping network cache update for instance because it is Building. {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 1057.312378] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Skipping network cache update for instance because it is Building. {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 1057.320774] env[61868]: DEBUG oslo_concurrency.lockutils [None req-fd7d4ac0-445f-48d4-98fe-3fdc7b61e43c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "interface-69a0b94b-2bf4-49ad-8fb0-ea945176e5eb-0bbbc97a-2c41-4d5d-a093-fca57a1d543b" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.004s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.346630] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "refresh_cache-a96cab5a-c053-4e45-96f6-2aba0a819110" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.346782] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquired lock "refresh_cache-a96cab5a-c053-4e45-96f6-2aba0a819110" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.347299] env[61868]: DEBUG nova.network.neutron [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Forcefully refreshing network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1057.347299] env[61868]: DEBUG nova.objects.instance [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lazy-loading 'info_cache' on Instance uuid a96cab5a-c053-4e45-96f6-2aba0a819110 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1057.403416] env[61868]: DEBUG oslo_vmware.api [None req-bd363ea7-bc29-46fa-adb7-999d56da820c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315958, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.760339} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.403679] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd363ea7-bc29-46fa-adb7-999d56da820c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1057.403866] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-bd363ea7-bc29-46fa-adb7-999d56da820c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Deleted contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1057.404110] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-bd363ea7-bc29-46fa-adb7-999d56da820c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1057.404363] env[61868]: INFO nova.compute.manager [None req-bd363ea7-bc29-46fa-adb7-999d56da820c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Took 1.63 seconds to destroy the instance on the hypervisor. [ 1057.404613] env[61868]: DEBUG oslo.service.loopingcall [None req-bd363ea7-bc29-46fa-adb7-999d56da820c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1057.404802] env[61868]: DEBUG nova.compute.manager [-] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1057.404897] env[61868]: DEBUG nova.network.neutron [-] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1057.610685] env[61868]: DEBUG oslo_vmware.api [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Task: {'id': task-1315961, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.708883] env[61868]: DEBUG oslo_vmware.api [None req-9010f905-ecb4-4531-ab1a-1061c24fb7bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315962, 'name': Destroy_Task} progress is 33%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.808076] env[61868]: DEBUG nova.network.neutron [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1057.826923] env[61868]: DEBUG oslo_concurrency.lockutils [None req-6ff3b3b2-959e-43b8-91b8-a6664466dcd8 tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "3a94eb16-cce3-4765-bc61-b6776bdd5b63" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.552s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.984589] env[61868]: DEBUG nova.network.neutron [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Updating instance_info_cache with network_info: [{"id": "a34b4784-533a-42f9-b5c9-ea8332e8eecb", "address": "fa:16:3e:18:b9:09", "network": {"id": "3535e048-fe57-4cbc-9489-0b77484069bd", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-748518849-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c8440ce785c344d793dd513b5f008d65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa34b4784-53", "ovs_interfaceid": "a34b4784-533a-42f9-b5c9-ea8332e8eecb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.111022] env[61868]: DEBUG oslo_vmware.api [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Task: {'id': task-1315961, 'name': ReconfigVM_Task, 'duration_secs': 0.723913} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.111354] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Reconfigured VM instance instance-00000065 to attach disk [datastore1] 1cb04c98-f725-4b1f-933e-ded580d340fb/1cb04c98-f725-4b1f-933e-ded580d340fb.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1058.111955] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-500e9aa4-732c-4e4e-b11a-2526e6e9ebb6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.118252] env[61868]: DEBUG oslo_vmware.api [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Waiting for the task: (returnval){ [ 1058.118252] env[61868]: value = "task-1315963" [ 1058.118252] env[61868]: _type = "Task" [ 1058.118252] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.125484] env[61868]: DEBUG oslo_vmware.api [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Task: {'id': task-1315963, 'name': Rename_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.192122] env[61868]: DEBUG nova.compute.manager [req-3460f400-99c4-4485-bd82-099c8a180a57 req-f33cc110-a7e2-48e2-ab0f-6233b5678fea service nova] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Received event network-vif-deleted-f8f4c521-03cd-4e54-99f1-761d9a515aa3 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1058.192361] env[61868]: INFO nova.compute.manager [req-3460f400-99c4-4485-bd82-099c8a180a57 req-f33cc110-a7e2-48e2-ab0f-6233b5678fea service nova] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Neutron deleted interface f8f4c521-03cd-4e54-99f1-761d9a515aa3; detaching it from the instance and deleting it from the info cache [ 1058.192508] env[61868]: DEBUG nova.network.neutron [req-3460f400-99c4-4485-bd82-099c8a180a57 req-f33cc110-a7e2-48e2-ab0f-6233b5678fea service nova] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.209890] env[61868]: DEBUG oslo_vmware.api [None req-9010f905-ecb4-4531-ab1a-1061c24fb7bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315962, 'name': Destroy_Task, 'duration_secs': 0.578987} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.210250] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-9010f905-ecb4-4531-ab1a-1061c24fb7bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Destroyed the VM [ 1058.211116] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9010f905-ecb4-4531-ab1a-1061c24fb7bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Deleting Snapshot of the VM instance {{(pid=61868) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1058.211116] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-ba90ce0e-fa1c-4b53-b2bd-ae595e98dc54 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.219849] env[61868]: DEBUG oslo_vmware.api [None req-9010f905-ecb4-4531-ab1a-1061c24fb7bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 1058.219849] env[61868]: value = "task-1315964" [ 1058.219849] env[61868]: _type = "Task" [ 1058.219849] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.227959] env[61868]: DEBUG oslo_vmware.api [None req-9010f905-ecb4-4531-ab1a-1061c24fb7bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315964, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.487986] env[61868]: DEBUG oslo_concurrency.lockutils [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Releasing lock "refresh_cache-273a38e7-8eb0-46fc-b9e2-31a9a7fc34db" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.488149] env[61868]: DEBUG nova.compute.manager [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Instance network_info: |[{"id": "a34b4784-533a-42f9-b5c9-ea8332e8eecb", "address": "fa:16:3e:18:b9:09", "network": {"id": "3535e048-fe57-4cbc-9489-0b77484069bd", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-748518849-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c8440ce785c344d793dd513b5f008d65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa34b4784-53", "ovs_interfaceid": "a34b4784-533a-42f9-b5c9-ea8332e8eecb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1058.488579] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:18:b9:09', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'da0e5087-d65b-416f-90fe-beaa9c534ad3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a34b4784-533a-42f9-b5c9-ea8332e8eecb', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1058.496167] env[61868]: DEBUG oslo.service.loopingcall [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1058.496436] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1058.496678] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d070a333-6284-48a9-99fa-ad2247b3f06e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.518564] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1058.518564] env[61868]: value = "task-1315965" [ 1058.518564] env[61868]: _type = "Task" [ 1058.518564] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.529014] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315965, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.537510] env[61868]: DEBUG nova.network.neutron [-] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.628651] env[61868]: DEBUG oslo_vmware.api [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Task: {'id': task-1315963, 'name': Rename_Task, 'duration_secs': 0.140693} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.628942] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1058.629212] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d0d1ca3b-da48-43eb-8e17-3f7c98144772 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.635574] env[61868]: DEBUG oslo_vmware.api [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Waiting for the task: (returnval){ [ 1058.635574] env[61868]: value = "task-1315966" [ 1058.635574] env[61868]: _type = "Task" [ 1058.635574] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.643746] env[61868]: DEBUG oslo_vmware.api [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Task: {'id': task-1315966, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.695203] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-71602a8c-864c-4e44-bd5e-f787a25268f9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.705053] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a161e3b8-4718-4821-9c6f-6996dbb9f79d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.731851] env[61868]: DEBUG nova.compute.manager [req-3460f400-99c4-4485-bd82-099c8a180a57 req-f33cc110-a7e2-48e2-ab0f-6233b5678fea service nova] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Detach interface failed, port_id=f8f4c521-03cd-4e54-99f1-761d9a515aa3, reason: Instance 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb could not be found. {{(pid=61868) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1058.735342] env[61868]: DEBUG oslo_vmware.api [None req-9010f905-ecb4-4531-ab1a-1061c24fb7bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315964, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.029737] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315965, 'name': CreateVM_Task, 'duration_secs': 0.36461} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.029943] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1059.030703] env[61868]: DEBUG oslo_concurrency.lockutils [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.030874] env[61868]: DEBUG oslo_concurrency.lockutils [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.031272] env[61868]: DEBUG oslo_concurrency.lockutils [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1059.031509] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fbb1ece5-b927-43e8-8b76-64d51e781298 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.039161] env[61868]: DEBUG oslo_vmware.api [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 1059.039161] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5281179c-8924-a41d-035b-807aec3b2e26" [ 1059.039161] env[61868]: _type = "Task" [ 1059.039161] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.043556] env[61868]: INFO nova.compute.manager [-] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Took 1.64 seconds to deallocate network for instance. [ 1059.052936] env[61868]: DEBUG oslo_vmware.api [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5281179c-8924-a41d-035b-807aec3b2e26, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.075659] env[61868]: DEBUG nova.network.neutron [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Updating instance_info_cache with network_info: [{"id": "cd9efdc6-e6f5-4449-b95d-d1fbec53a295", "address": "fa:16:3e:bf:13:19", "network": {"id": "d4f100cd-8de6-4459-a7a6-bc1895b8b14f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-404440555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7fc76299baf4a97b57139e5f1caa16e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd9efdc6-e6", "ovs_interfaceid": "cd9efdc6-e6f5-4449-b95d-d1fbec53a295", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1059.145689] env[61868]: DEBUG oslo_vmware.api [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Task: {'id': task-1315966, 'name': PowerOnVM_Task, 'duration_secs': 0.441006} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.145947] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1059.146169] env[61868]: INFO nova.compute.manager [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Took 7.48 seconds to spawn the instance on the hypervisor. [ 1059.146394] env[61868]: DEBUG nova.compute.manager [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1059.147152] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2355e835-603a-43a0-ac61-ed9446f3bbdb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.229909] env[61868]: DEBUG oslo_vmware.api [None req-9010f905-ecb4-4531-ab1a-1061c24fb7bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315964, 'name': RemoveSnapshot_Task, 'duration_secs': 0.524437} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.230323] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-9010f905-ecb4-4531-ab1a-1061c24fb7bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Deleted Snapshot of the VM instance {{(pid=61868) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1059.299849] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0f1c9ffc-1cdf-4cea-a6d7-72f2469e0b5f tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "3a94eb16-cce3-4765-bc61-b6776bdd5b63" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.300132] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0f1c9ffc-1cdf-4cea-a6d7-72f2469e0b5f tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "3a94eb16-cce3-4765-bc61-b6776bdd5b63" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.300349] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0f1c9ffc-1cdf-4cea-a6d7-72f2469e0b5f tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "3a94eb16-cce3-4765-bc61-b6776bdd5b63-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.300536] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0f1c9ffc-1cdf-4cea-a6d7-72f2469e0b5f tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "3a94eb16-cce3-4765-bc61-b6776bdd5b63-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.300709] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0f1c9ffc-1cdf-4cea-a6d7-72f2469e0b5f tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "3a94eb16-cce3-4765-bc61-b6776bdd5b63-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.302664] env[61868]: INFO nova.compute.manager [None req-0f1c9ffc-1cdf-4cea-a6d7-72f2469e0b5f tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Terminating instance [ 1059.304544] env[61868]: DEBUG nova.compute.manager [None req-0f1c9ffc-1cdf-4cea-a6d7-72f2469e0b5f tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1059.304741] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0f1c9ffc-1cdf-4cea-a6d7-72f2469e0b5f tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1059.305553] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd7d9113-13a5-44d6-931b-402c427a39a0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.312780] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0f1c9ffc-1cdf-4cea-a6d7-72f2469e0b5f tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1059.313009] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b38a84d6-32ea-4115-a66e-6a04b2db04a6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.379727] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0f1c9ffc-1cdf-4cea-a6d7-72f2469e0b5f tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1059.379969] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0f1c9ffc-1cdf-4cea-a6d7-72f2469e0b5f tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Deleting contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1059.380178] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f1c9ffc-1cdf-4cea-a6d7-72f2469e0b5f tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Deleting the datastore file [datastore2] 3a94eb16-cce3-4765-bc61-b6776bdd5b63 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1059.380435] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0036c945-cb9c-46cd-a5ba-e2bb1531a14b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.386061] env[61868]: DEBUG oslo_vmware.api [None req-0f1c9ffc-1cdf-4cea-a6d7-72f2469e0b5f tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for the task: (returnval){ [ 1059.386061] env[61868]: value = "task-1315968" [ 1059.386061] env[61868]: _type = "Task" [ 1059.386061] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.393916] env[61868]: DEBUG oslo_vmware.api [None req-0f1c9ffc-1cdf-4cea-a6d7-72f2469e0b5f tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315968, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.548840] env[61868]: DEBUG oslo_vmware.api [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5281179c-8924-a41d-035b-807aec3b2e26, 'name': SearchDatastore_Task, 'duration_secs': 0.01737} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.549164] env[61868]: DEBUG oslo_concurrency.lockutils [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.549405] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1059.549640] env[61868]: DEBUG oslo_concurrency.lockutils [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.549791] env[61868]: DEBUG oslo_concurrency.lockutils [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.549973] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1059.550245] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-db7b27d1-032c-4b28-a574-b8fc7c140330 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.558896] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bd363ea7-bc29-46fa-adb7-999d56da820c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.559039] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bd363ea7-bc29-46fa-adb7-999d56da820c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.559280] env[61868]: DEBUG nova.objects.instance [None req-bd363ea7-bc29-46fa-adb7-999d56da820c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lazy-loading 'resources' on Instance uuid 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1059.560275] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1059.560456] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1059.561360] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1361046c-5d45-4bb6-9891-cd4853689efa {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.566606] env[61868]: DEBUG oslo_vmware.api [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 1059.566606] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52395ffd-8d5e-3d0c-ee2a-c280752c9a22" [ 1059.566606] env[61868]: _type = "Task" [ 1059.566606] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.574509] env[61868]: DEBUG oslo_vmware.api [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52395ffd-8d5e-3d0c-ee2a-c280752c9a22, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.577984] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Releasing lock "refresh_cache-a96cab5a-c053-4e45-96f6-2aba0a819110" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.578210] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Updated the network info_cache for instance {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1059.578394] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1059.578788] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1059.578950] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1059.579123] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1059.579272] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1059.579417] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1059.579545] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61868) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1059.579689] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager.update_available_resource {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1059.661904] env[61868]: INFO nova.compute.manager [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Took 12.21 seconds to build instance. [ 1059.736999] env[61868]: WARNING nova.compute.manager [None req-9010f905-ecb4-4531-ab1a-1061c24fb7bb tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Image not found during snapshot: nova.exception.ImageNotFound: Image ee4de7b2-a1b2-473c-b4f8-a9e13795fe3c could not be found. [ 1059.895749] env[61868]: DEBUG oslo_vmware.api [None req-0f1c9ffc-1cdf-4cea-a6d7-72f2469e0b5f tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Task: {'id': task-1315968, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.123508} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.896068] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f1c9ffc-1cdf-4cea-a6d7-72f2469e0b5f tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1059.896266] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0f1c9ffc-1cdf-4cea-a6d7-72f2469e0b5f tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Deleted contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1059.896453] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0f1c9ffc-1cdf-4cea-a6d7-72f2469e0b5f tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1059.896637] env[61868]: INFO nova.compute.manager [None req-0f1c9ffc-1cdf-4cea-a6d7-72f2469e0b5f tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Took 0.59 seconds to destroy the instance on the hypervisor. [ 1059.896883] env[61868]: DEBUG oslo.service.loopingcall [None req-0f1c9ffc-1cdf-4cea-a6d7-72f2469e0b5f tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1059.897098] env[61868]: DEBUG nova.compute.manager [-] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1059.897199] env[61868]: DEBUG nova.network.neutron [-] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1060.076561] env[61868]: DEBUG oslo_vmware.api [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52395ffd-8d5e-3d0c-ee2a-c280752c9a22, 'name': SearchDatastore_Task, 'duration_secs': 0.008105} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.077379] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e895aed-682a-4651-b63f-e836b79da4f6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.081999] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.083469] env[61868]: DEBUG oslo_vmware.api [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 1060.083469] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52f9ad0d-f557-3844-e169-31f6c0d71ec0" [ 1060.083469] env[61868]: _type = "Task" [ 1060.083469] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.091146] env[61868]: DEBUG oslo_vmware.api [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52f9ad0d-f557-3844-e169-31f6c0d71ec0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.114853] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d92a9505-83f8-4eae-a808-e732088fb928 tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Acquiring lock "1cb04c98-f725-4b1f-933e-ded580d340fb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.163635] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7c86cf27-5ba2-4826-a7b7-aa832dcec5dc tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Lock "1cb04c98-f725-4b1f-933e-ded580d340fb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.722s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.163985] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d92a9505-83f8-4eae-a808-e732088fb928 tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Lock "1cb04c98-f725-4b1f-933e-ded580d340fb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.049s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.164333] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d92a9505-83f8-4eae-a808-e732088fb928 tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Acquiring lock "1cb04c98-f725-4b1f-933e-ded580d340fb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.164551] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d92a9505-83f8-4eae-a808-e732088fb928 tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Lock "1cb04c98-f725-4b1f-933e-ded580d340fb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.164729] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d92a9505-83f8-4eae-a808-e732088fb928 tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Lock "1cb04c98-f725-4b1f-933e-ded580d340fb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.166825] env[61868]: INFO nova.compute.manager [None req-d92a9505-83f8-4eae-a808-e732088fb928 tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Terminating instance [ 1060.169064] env[61868]: DEBUG nova.compute.manager [None req-d92a9505-83f8-4eae-a808-e732088fb928 tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1060.169266] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-d92a9505-83f8-4eae-a808-e732088fb928 tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1060.170345] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef057880-81b5-44d0-97dc-f07c47eb447f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.174224] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-677ce3e6-5958-4d28-99c2-1322dc0ec931 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.182962] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20c56bd6-6579-4576-b2f0-12e216c73905 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.186608] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-d92a9505-83f8-4eae-a808-e732088fb928 tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1060.186608] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b88c47d3-0b1e-42ae-b74a-7c5fe1516ef6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.218695] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbf49f38-7e92-462f-b110-1ba64482ceeb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.221798] env[61868]: DEBUG oslo_vmware.api [None req-d92a9505-83f8-4eae-a808-e732088fb928 tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Waiting for the task: (returnval){ [ 1060.221798] env[61868]: value = "task-1315969" [ 1060.221798] env[61868]: _type = "Task" [ 1060.221798] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.224017] env[61868]: DEBUG nova.compute.manager [req-b072138e-021d-48bb-a59a-0595ff4b2477 req-6f38ab62-8585-4b50-bac0-59465dac1f48 service nova] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Received event network-vif-deleted-32232776-6205-48a9-abcc-b24add5b69a1 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1060.224257] env[61868]: INFO nova.compute.manager [req-b072138e-021d-48bb-a59a-0595ff4b2477 req-6f38ab62-8585-4b50-bac0-59465dac1f48 service nova] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Neutron deleted interface 32232776-6205-48a9-abcc-b24add5b69a1; detaching it from the instance and deleting it from the info cache [ 1060.224470] env[61868]: DEBUG nova.network.neutron [req-b072138e-021d-48bb-a59a-0595ff4b2477 req-6f38ab62-8585-4b50-bac0-59465dac1f48 service nova] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.232840] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5157fe53-44b8-48e3-9a1f-d93b833c25c5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.240659] env[61868]: DEBUG oslo_vmware.api [None req-d92a9505-83f8-4eae-a808-e732088fb928 tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Task: {'id': task-1315969, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.250962] env[61868]: DEBUG nova.compute.provider_tree [None req-bd363ea7-bc29-46fa-adb7-999d56da820c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1060.593370] env[61868]: DEBUG oslo_vmware.api [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52f9ad0d-f557-3844-e169-31f6c0d71ec0, 'name': SearchDatastore_Task, 'duration_secs': 0.016302} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.593692] env[61868]: DEBUG oslo_concurrency.lockutils [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.593960] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db/273a38e7-8eb0-46fc-b9e2-31a9a7fc34db.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1060.594301] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e97b80fb-7049-4ac3-9eb3-027850c2e5dd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.601114] env[61868]: DEBUG oslo_vmware.api [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 1060.601114] env[61868]: value = "task-1315970" [ 1060.601114] env[61868]: _type = "Task" [ 1060.601114] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.608370] env[61868]: DEBUG oslo_vmware.api [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315970, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.626780] env[61868]: DEBUG nova.network.neutron [-] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.732527] env[61868]: DEBUG oslo_vmware.api [None req-d92a9505-83f8-4eae-a808-e732088fb928 tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Task: {'id': task-1315969, 'name': PowerOffVM_Task, 'duration_secs': 0.170163} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.732785] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-40b7ea06-be89-44b8-9d75-72960aeb40b4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.734756] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-d92a9505-83f8-4eae-a808-e732088fb928 tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1060.734947] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-d92a9505-83f8-4eae-a808-e732088fb928 tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1060.735221] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ee8f92e3-33c3-42a4-8237-269e58d0f55a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.743618] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0349fa98-5cff-4ae8-acf5-3843e330d512 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.755056] env[61868]: DEBUG nova.scheduler.client.report [None req-bd363ea7-bc29-46fa-adb7-999d56da820c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1060.773260] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eedd44b4-3d1a-456b-a399-aa2e568c2d72 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.773556] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eedd44b4-3d1a-456b-a399-aa2e568c2d72 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.773775] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eedd44b4-3d1a-456b-a399-aa2e568c2d72 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.773964] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eedd44b4-3d1a-456b-a399-aa2e568c2d72 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.774163] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eedd44b4-3d1a-456b-a399-aa2e568c2d72 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.775922] env[61868]: DEBUG nova.compute.manager [req-b072138e-021d-48bb-a59a-0595ff4b2477 req-6f38ab62-8585-4b50-bac0-59465dac1f48 service nova] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Detach interface failed, port_id=32232776-6205-48a9-abcc-b24add5b69a1, reason: Instance 3a94eb16-cce3-4765-bc61-b6776bdd5b63 could not be found. {{(pid=61868) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1060.776688] env[61868]: INFO nova.compute.manager [None req-eedd44b4-3d1a-456b-a399-aa2e568c2d72 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Terminating instance [ 1060.778673] env[61868]: DEBUG nova.compute.manager [None req-eedd44b4-3d1a-456b-a399-aa2e568c2d72 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1060.779225] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-eedd44b4-3d1a-456b-a399-aa2e568c2d72 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1060.779687] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6b43d72-90fb-4a75-be6d-abdcde70daef {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.787273] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-eedd44b4-3d1a-456b-a399-aa2e568c2d72 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1060.787782] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-36f0a97f-8baa-4f6c-b126-f00b0646737c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.794901] env[61868]: DEBUG oslo_vmware.api [None req-eedd44b4-3d1a-456b-a399-aa2e568c2d72 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 1060.794901] env[61868]: value = "task-1315972" [ 1060.794901] env[61868]: _type = "Task" [ 1060.794901] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.803639] env[61868]: DEBUG oslo_vmware.api [None req-eedd44b4-3d1a-456b-a399-aa2e568c2d72 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315972, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.926191] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-d92a9505-83f8-4eae-a808-e732088fb928 tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1060.926551] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-d92a9505-83f8-4eae-a808-e732088fb928 tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1060.926772] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-d92a9505-83f8-4eae-a808-e732088fb928 tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Deleting the datastore file [datastore1] 1cb04c98-f725-4b1f-933e-ded580d340fb {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1060.927077] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-87c21b8d-d4b5-4ade-821c-645c60c2ef43 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.935760] env[61868]: DEBUG oslo_vmware.api [None req-d92a9505-83f8-4eae-a808-e732088fb928 tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Waiting for the task: (returnval){ [ 1060.935760] env[61868]: value = "task-1315973" [ 1060.935760] env[61868]: _type = "Task" [ 1060.935760] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.945948] env[61868]: DEBUG oslo_vmware.api [None req-d92a9505-83f8-4eae-a808-e732088fb928 tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Task: {'id': task-1315973, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.111088] env[61868]: DEBUG oslo_vmware.api [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315970, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.44913} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.111394] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db/273a38e7-8eb0-46fc-b9e2-31a9a7fc34db.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1061.111612] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1061.111866] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a6db4a0f-dbdc-4f28-806e-947ea9a530df {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.117539] env[61868]: DEBUG oslo_vmware.api [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 1061.117539] env[61868]: value = "task-1315974" [ 1061.117539] env[61868]: _type = "Task" [ 1061.117539] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.124944] env[61868]: DEBUG oslo_vmware.api [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315974, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.129421] env[61868]: INFO nova.compute.manager [-] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Took 1.23 seconds to deallocate network for instance. [ 1061.259802] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bd363ea7-bc29-46fa-adb7-999d56da820c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.700s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.261995] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 1.180s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.262190] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.262392] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61868) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1061.263270] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbd98b87-139b-416f-bce3-b8c88961d299 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.271414] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a43b4d49-15a4-4671-b9a5-42ced8e2511e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.285857] env[61868]: INFO nova.scheduler.client.report [None req-bd363ea7-bc29-46fa-adb7-999d56da820c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Deleted allocations for instance 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb [ 1061.287408] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a24b60c2-db46-4a44-986e-27b303cdec09 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.298796] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac93b150-fee4-448e-b5f5-538b2f4f900f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.306588] env[61868]: DEBUG oslo_vmware.api [None req-eedd44b4-3d1a-456b-a399-aa2e568c2d72 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315972, 'name': PowerOffVM_Task, 'duration_secs': 0.316102} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.329412] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-eedd44b4-3d1a-456b-a399-aa2e568c2d72 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1061.329607] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-eedd44b4-3d1a-456b-a399-aa2e568c2d72 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1061.329917] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180164MB free_disk=134GB free_vcpus=48 pci_devices=None {{(pid=61868) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1061.330063] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.330252] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.332173] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4926e336-d06b-42ab-b9bf-d1892b4f608b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.433985] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-eedd44b4-3d1a-456b-a399-aa2e568c2d72 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1061.434290] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-eedd44b4-3d1a-456b-a399-aa2e568c2d72 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Deleting contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1061.434502] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-eedd44b4-3d1a-456b-a399-aa2e568c2d72 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Deleting the datastore file [datastore2] 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1061.434762] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ece580e9-a24e-48fe-a5c8-b0604f3fcd17 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.444739] env[61868]: DEBUG oslo_vmware.api [None req-d92a9505-83f8-4eae-a808-e732088fb928 tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Task: {'id': task-1315973, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.229506} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.445834] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-d92a9505-83f8-4eae-a808-e732088fb928 tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1061.446046] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-d92a9505-83f8-4eae-a808-e732088fb928 tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1061.446239] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-d92a9505-83f8-4eae-a808-e732088fb928 tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1061.446436] env[61868]: INFO nova.compute.manager [None req-d92a9505-83f8-4eae-a808-e732088fb928 tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Took 1.28 seconds to destroy the instance on the hypervisor. [ 1061.446721] env[61868]: DEBUG oslo.service.loopingcall [None req-d92a9505-83f8-4eae-a808-e732088fb928 tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1061.447012] env[61868]: DEBUG oslo_vmware.api [None req-eedd44b4-3d1a-456b-a399-aa2e568c2d72 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for the task: (returnval){ [ 1061.447012] env[61868]: value = "task-1315976" [ 1061.447012] env[61868]: _type = "Task" [ 1061.447012] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.447221] env[61868]: DEBUG nova.compute.manager [-] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1061.447320] env[61868]: DEBUG nova.network.neutron [-] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1061.456632] env[61868]: DEBUG oslo_vmware.api [None req-eedd44b4-3d1a-456b-a399-aa2e568c2d72 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315976, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.629461] env[61868]: DEBUG oslo_vmware.api [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315974, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076407} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.629744] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1061.630501] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eb601f7-7565-4f1e-b2bb-591ec4b7d5dd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.644153] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0f1c9ffc-1cdf-4cea-a6d7-72f2469e0b5f tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.652705] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db/273a38e7-8eb0-46fc-b9e2-31a9a7fc34db.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1061.652963] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7456cd98-ae31-43eb-9ab6-4ecc4004d8f8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.673680] env[61868]: DEBUG oslo_vmware.api [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 1061.673680] env[61868]: value = "task-1315977" [ 1061.673680] env[61868]: _type = "Task" [ 1061.673680] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.682201] env[61868]: DEBUG oslo_vmware.api [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315977, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.795567] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bd363ea7-bc29-46fa-adb7-999d56da820c tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "69a0b94b-2bf4-49ad-8fb0-ea945176e5eb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.034s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.958426] env[61868]: DEBUG oslo_vmware.api [None req-eedd44b4-3d1a-456b-a399-aa2e568c2d72 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Task: {'id': task-1315976, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.1358} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.958688] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-eedd44b4-3d1a-456b-a399-aa2e568c2d72 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1061.958993] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-eedd44b4-3d1a-456b-a399-aa2e568c2d72 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Deleted contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1061.959078] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-eedd44b4-3d1a-456b-a399-aa2e568c2d72 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1061.959241] env[61868]: INFO nova.compute.manager [None req-eedd44b4-3d1a-456b-a399-aa2e568c2d72 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Took 1.18 seconds to destroy the instance on the hypervisor. [ 1061.959484] env[61868]: DEBUG oslo.service.loopingcall [None req-eedd44b4-3d1a-456b-a399-aa2e568c2d72 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1061.959837] env[61868]: DEBUG nova.compute.manager [-] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1061.959969] env[61868]: DEBUG nova.network.neutron [-] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1062.179237] env[61868]: DEBUG nova.network.neutron [-] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.186692] env[61868]: DEBUG oslo_vmware.api [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315977, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.250508] env[61868]: DEBUG nova.compute.manager [req-ae175c82-4cf2-4662-b70c-c61e70e50a1c req-5e099b9e-6f25-47aa-a462-7cc997b942e2 service nova] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Received event network-vif-deleted-0d4ba939-0509-4e44-885e-2684970089d3 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1062.250720] env[61868]: DEBUG nova.compute.manager [req-ae175c82-4cf2-4662-b70c-c61e70e50a1c req-5e099b9e-6f25-47aa-a462-7cc997b942e2 service nova] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Received event network-vif-deleted-89658ae1-cd68-4ab5-9ddd-7de554905b00 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1062.250889] env[61868]: INFO nova.compute.manager [req-ae175c82-4cf2-4662-b70c-c61e70e50a1c req-5e099b9e-6f25-47aa-a462-7cc997b942e2 service nova] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Neutron deleted interface 89658ae1-cd68-4ab5-9ddd-7de554905b00; detaching it from the instance and deleting it from the info cache [ 1062.251073] env[61868]: DEBUG nova.network.neutron [req-ae175c82-4cf2-4662-b70c-c61e70e50a1c req-5e099b9e-6f25-47aa-a462-7cc997b942e2 service nova] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.354331] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance a96cab5a-c053-4e45-96f6-2aba0a819110 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1062.354500] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1062.354651] env[61868]: WARNING nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 3a94eb16-cce3-4765-bc61-b6776bdd5b63 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1062.354796] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 1cb04c98-f725-4b1f-933e-ded580d340fb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1062.354918] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1062.355108] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=61868) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1062.355248] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=61868) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1062.427212] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9020bfcc-6fd6-4118-9656-deefe9dfe428 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.434807] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3c65ebf-1724-498f-86a4-4222940013f6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.465380] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5565ba4f-5c8c-40f1-b6f9-ed4f74b92ce1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.472194] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9caa5815-84bc-406b-a275-2115365ab472 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.485145] env[61868]: DEBUG nova.compute.provider_tree [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1062.514970] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da0b0370-9019-46e0-b203-afacf721d587 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "a96cab5a-c053-4e45-96f6-2aba0a819110" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.515219] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da0b0370-9019-46e0-b203-afacf721d587 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "a96cab5a-c053-4e45-96f6-2aba0a819110" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.515459] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da0b0370-9019-46e0-b203-afacf721d587 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "a96cab5a-c053-4e45-96f6-2aba0a819110-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.515615] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da0b0370-9019-46e0-b203-afacf721d587 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "a96cab5a-c053-4e45-96f6-2aba0a819110-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.515781] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da0b0370-9019-46e0-b203-afacf721d587 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "a96cab5a-c053-4e45-96f6-2aba0a819110-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.517655] env[61868]: INFO nova.compute.manager [None req-da0b0370-9019-46e0-b203-afacf721d587 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Terminating instance [ 1062.519380] env[61868]: DEBUG nova.compute.manager [None req-da0b0370-9019-46e0-b203-afacf721d587 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1062.519580] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-da0b0370-9019-46e0-b203-afacf721d587 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1062.520326] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e0b268a-aa59-4b81-91b7-af432ef0db01 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.527678] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-da0b0370-9019-46e0-b203-afacf721d587 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1062.527901] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8a8a5dd2-9fb8-411c-94d0-f566dc997231 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.533650] env[61868]: DEBUG oslo_vmware.api [None req-da0b0370-9019-46e0-b203-afacf721d587 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for the task: (returnval){ [ 1062.533650] env[61868]: value = "task-1315978" [ 1062.533650] env[61868]: _type = "Task" [ 1062.533650] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.541349] env[61868]: DEBUG oslo_vmware.api [None req-da0b0370-9019-46e0-b203-afacf721d587 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315978, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.684420] env[61868]: INFO nova.compute.manager [-] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Took 1.24 seconds to deallocate network for instance. [ 1062.684764] env[61868]: DEBUG nova.network.neutron [-] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.686526] env[61868]: DEBUG oslo_vmware.api [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315977, 'name': ReconfigVM_Task, 'duration_secs': 0.817317} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.688310] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Reconfigured VM instance instance-00000066 to attach disk [datastore1] 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db/273a38e7-8eb0-46fc-b9e2-31a9a7fc34db.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1062.691319] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ce62bb90-1be8-4e9a-bd96-8b329809af4c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.697971] env[61868]: DEBUG oslo_vmware.api [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 1062.697971] env[61868]: value = "task-1315979" [ 1062.697971] env[61868]: _type = "Task" [ 1062.697971] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.710846] env[61868]: DEBUG oslo_vmware.api [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315979, 'name': Rename_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.753389] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-20ce5ea6-f23a-4b98-874d-359c97f8638a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.766674] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8202e703-58ae-4e29-bfb2-6101c68ce0db {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.797435] env[61868]: DEBUG nova.compute.manager [req-ae175c82-4cf2-4662-b70c-c61e70e50a1c req-5e099b9e-6f25-47aa-a462-7cc997b942e2 service nova] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Detach interface failed, port_id=89658ae1-cd68-4ab5-9ddd-7de554905b00, reason: Instance 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1 could not be found. {{(pid=61868) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1062.988659] env[61868]: DEBUG nova.scheduler.client.report [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1063.043833] env[61868]: DEBUG oslo_vmware.api [None req-da0b0370-9019-46e0-b203-afacf721d587 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315978, 'name': PowerOffVM_Task, 'duration_secs': 0.18783} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.044104] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-da0b0370-9019-46e0-b203-afacf721d587 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1063.044303] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-da0b0370-9019-46e0-b203-afacf721d587 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1063.044555] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6041c5c2-1d9d-442c-ac03-d95050539141 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.103219] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-da0b0370-9019-46e0-b203-afacf721d587 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1063.103462] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-da0b0370-9019-46e0-b203-afacf721d587 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1063.103647] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-da0b0370-9019-46e0-b203-afacf721d587 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Deleting the datastore file [datastore1] a96cab5a-c053-4e45-96f6-2aba0a819110 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1063.103938] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d8e41b75-6617-4f0f-933a-f75129b38942 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.110727] env[61868]: DEBUG oslo_vmware.api [None req-da0b0370-9019-46e0-b203-afacf721d587 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for the task: (returnval){ [ 1063.110727] env[61868]: value = "task-1315981" [ 1063.110727] env[61868]: _type = "Task" [ 1063.110727] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.118508] env[61868]: DEBUG oslo_vmware.api [None req-da0b0370-9019-46e0-b203-afacf721d587 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315981, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.187572] env[61868]: INFO nova.compute.manager [-] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Took 1.23 seconds to deallocate network for instance. [ 1063.194957] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d92a9505-83f8-4eae-a808-e732088fb928 tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.208769] env[61868]: DEBUG oslo_vmware.api [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315979, 'name': Rename_Task, 'duration_secs': 0.128803} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.208769] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1063.208932] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1f9f30b0-064e-46f7-bddb-2a4a15e3498f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.216529] env[61868]: DEBUG oslo_vmware.api [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 1063.216529] env[61868]: value = "task-1315982" [ 1063.216529] env[61868]: _type = "Task" [ 1063.216529] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.224472] env[61868]: DEBUG oslo_vmware.api [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315982, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.493795] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61868) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1063.493973] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.164s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.494332] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0f1c9ffc-1cdf-4cea-a6d7-72f2469e0b5f tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.850s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.494567] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0f1c9ffc-1cdf-4cea-a6d7-72f2469e0b5f tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.496937] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d92a9505-83f8-4eae-a808-e732088fb928 tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.302s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.497179] env[61868]: DEBUG nova.objects.instance [None req-d92a9505-83f8-4eae-a808-e732088fb928 tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Lazy-loading 'resources' on Instance uuid 1cb04c98-f725-4b1f-933e-ded580d340fb {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1063.515933] env[61868]: INFO nova.scheduler.client.report [None req-0f1c9ffc-1cdf-4cea-a6d7-72f2469e0b5f tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Deleted allocations for instance 3a94eb16-cce3-4765-bc61-b6776bdd5b63 [ 1063.620545] env[61868]: DEBUG oslo_vmware.api [None req-da0b0370-9019-46e0-b203-afacf721d587 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Task: {'id': task-1315981, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143736} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.620822] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-da0b0370-9019-46e0-b203-afacf721d587 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1063.621036] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-da0b0370-9019-46e0-b203-afacf721d587 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1063.621261] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-da0b0370-9019-46e0-b203-afacf721d587 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1063.621447] env[61868]: INFO nova.compute.manager [None req-da0b0370-9019-46e0-b203-afacf721d587 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1063.621705] env[61868]: DEBUG oslo.service.loopingcall [None req-da0b0370-9019-46e0-b203-afacf721d587 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1063.621910] env[61868]: DEBUG nova.compute.manager [-] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1063.622020] env[61868]: DEBUG nova.network.neutron [-] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1063.698240] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eedd44b4-3d1a-456b-a399-aa2e568c2d72 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.725824] env[61868]: DEBUG oslo_vmware.api [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1315982, 'name': PowerOnVM_Task, 'duration_secs': 0.433585} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.726129] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1063.726341] env[61868]: INFO nova.compute.manager [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Took 8.30 seconds to spawn the instance on the hypervisor. [ 1063.726527] env[61868]: DEBUG nova.compute.manager [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1063.727318] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5d3110c-14c8-4900-be14-d9193fdf19cb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.023645] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0f1c9ffc-1cdf-4cea-a6d7-72f2469e0b5f tempest-ServersTestJSON-136282107 tempest-ServersTestJSON-136282107-project-member] Lock "3a94eb16-cce3-4765-bc61-b6776bdd5b63" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.723s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.058910] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2943eb5b-dc5e-4bc7-a999-c991199ffd51 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.066924] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07d966ac-d719-4fdb-b14f-0b67de1a47c5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.097897] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4822c5d-76ca-4551-9198-f4a84d788cd0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.105127] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2540fd1-c75e-469f-b284-1ca71c026f22 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.118226] env[61868]: DEBUG nova.compute.provider_tree [None req-d92a9505-83f8-4eae-a808-e732088fb928 tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1064.245514] env[61868]: INFO nova.compute.manager [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Took 13.06 seconds to build instance. [ 1064.283922] env[61868]: DEBUG nova.compute.manager [req-842fa7a2-ef63-4e48-9a77-5034b8e24651 req-67d870f5-4530-4976-b7ad-686c94d478bb service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Received event network-vif-deleted-cd9efdc6-e6f5-4449-b95d-d1fbec53a295 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1064.284205] env[61868]: INFO nova.compute.manager [req-842fa7a2-ef63-4e48-9a77-5034b8e24651 req-67d870f5-4530-4976-b7ad-686c94d478bb service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Neutron deleted interface cd9efdc6-e6f5-4449-b95d-d1fbec53a295; detaching it from the instance and deleting it from the info cache [ 1064.284366] env[61868]: DEBUG nova.network.neutron [req-842fa7a2-ef63-4e48-9a77-5034b8e24651 req-67d870f5-4530-4976-b7ad-686c94d478bb service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1064.360561] env[61868]: DEBUG nova.network.neutron [-] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1064.621699] env[61868]: DEBUG nova.scheduler.client.report [None req-d92a9505-83f8-4eae-a808-e732088fb928 tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1064.698229] env[61868]: DEBUG nova.compute.manager [req-97ea538d-89b1-4d96-a0c7-8de9d5c32b38 req-bd501a0a-7af8-4f8d-8df0-f4046bdfc9cf service nova] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Received event network-changed-a34b4784-533a-42f9-b5c9-ea8332e8eecb {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1064.698853] env[61868]: DEBUG nova.compute.manager [req-97ea538d-89b1-4d96-a0c7-8de9d5c32b38 req-bd501a0a-7af8-4f8d-8df0-f4046bdfc9cf service nova] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Refreshing instance network info cache due to event network-changed-a34b4784-533a-42f9-b5c9-ea8332e8eecb. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1064.698853] env[61868]: DEBUG oslo_concurrency.lockutils [req-97ea538d-89b1-4d96-a0c7-8de9d5c32b38 req-bd501a0a-7af8-4f8d-8df0-f4046bdfc9cf service nova] Acquiring lock "refresh_cache-273a38e7-8eb0-46fc-b9e2-31a9a7fc34db" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.698853] env[61868]: DEBUG oslo_concurrency.lockutils [req-97ea538d-89b1-4d96-a0c7-8de9d5c32b38 req-bd501a0a-7af8-4f8d-8df0-f4046bdfc9cf service nova] Acquired lock "refresh_cache-273a38e7-8eb0-46fc-b9e2-31a9a7fc34db" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.699129] env[61868]: DEBUG nova.network.neutron [req-97ea538d-89b1-4d96-a0c7-8de9d5c32b38 req-bd501a0a-7af8-4f8d-8df0-f4046bdfc9cf service nova] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Refreshing network info cache for port a34b4784-533a-42f9-b5c9-ea8332e8eecb {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1064.748140] env[61868]: DEBUG oslo_concurrency.lockutils [None req-791e0e66-ee5d-46f4-97a3-146a2d84267e tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "273a38e7-8eb0-46fc-b9e2-31a9a7fc34db" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.572s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.786882] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-483e9c0d-0cdf-4538-b8b1-6c7d6b09637f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.800751] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-088833db-2f3f-497a-8041-835be545a37a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.824031] env[61868]: DEBUG nova.compute.manager [req-842fa7a2-ef63-4e48-9a77-5034b8e24651 req-67d870f5-4530-4976-b7ad-686c94d478bb service nova] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Detach interface failed, port_id=cd9efdc6-e6f5-4449-b95d-d1fbec53a295, reason: Instance a96cab5a-c053-4e45-96f6-2aba0a819110 could not be found. {{(pid=61868) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1064.863701] env[61868]: INFO nova.compute.manager [-] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Took 1.24 seconds to deallocate network for instance. [ 1065.126798] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d92a9505-83f8-4eae-a808-e732088fb928 tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.630s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.129045] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eedd44b4-3d1a-456b-a399-aa2e568c2d72 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.431s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.129288] env[61868]: DEBUG nova.objects.instance [None req-eedd44b4-3d1a-456b-a399-aa2e568c2d72 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lazy-loading 'resources' on Instance uuid 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1065.155263] env[61868]: INFO nova.scheduler.client.report [None req-d92a9505-83f8-4eae-a808-e732088fb928 tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Deleted allocations for instance 1cb04c98-f725-4b1f-933e-ded580d340fb [ 1065.370619] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da0b0370-9019-46e0-b203-afacf721d587 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.421739] env[61868]: DEBUG nova.network.neutron [req-97ea538d-89b1-4d96-a0c7-8de9d5c32b38 req-bd501a0a-7af8-4f8d-8df0-f4046bdfc9cf service nova] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Updated VIF entry in instance network info cache for port a34b4784-533a-42f9-b5c9-ea8332e8eecb. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1065.422135] env[61868]: DEBUG nova.network.neutron [req-97ea538d-89b1-4d96-a0c7-8de9d5c32b38 req-bd501a0a-7af8-4f8d-8df0-f4046bdfc9cf service nova] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Updating instance_info_cache with network_info: [{"id": "a34b4784-533a-42f9-b5c9-ea8332e8eecb", "address": "fa:16:3e:18:b9:09", "network": {"id": "3535e048-fe57-4cbc-9489-0b77484069bd", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-748518849-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c8440ce785c344d793dd513b5f008d65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa34b4784-53", "ovs_interfaceid": "a34b4784-533a-42f9-b5c9-ea8332e8eecb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1065.669354] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d92a9505-83f8-4eae-a808-e732088fb928 tempest-InstanceActionsV221TestJSON-1293783195 tempest-InstanceActionsV221TestJSON-1293783195-project-member] Lock "1cb04c98-f725-4b1f-933e-ded580d340fb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.505s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.694605] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eefd1a39-6043-498c-b38c-642431183ffb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.702992] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59527cc8-dbb1-4d7e-9256-96db995b2372 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.733526] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63f76032-5e43-4ebb-b593-e06d2ac4f50c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.741077] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90b57d35-6cff-4f02-8f2e-0c6db877cbcf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.755579] env[61868]: DEBUG nova.compute.provider_tree [None req-eedd44b4-3d1a-456b-a399-aa2e568c2d72 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1065.925397] env[61868]: DEBUG oslo_concurrency.lockutils [req-97ea538d-89b1-4d96-a0c7-8de9d5c32b38 req-bd501a0a-7af8-4f8d-8df0-f4046bdfc9cf service nova] Releasing lock "refresh_cache-273a38e7-8eb0-46fc-b9e2-31a9a7fc34db" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1066.259114] env[61868]: DEBUG nova.scheduler.client.report [None req-eedd44b4-3d1a-456b-a399-aa2e568c2d72 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1066.764476] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eedd44b4-3d1a-456b-a399-aa2e568c2d72 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.635s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.766926] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da0b0370-9019-46e0-b203-afacf721d587 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.396s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.767219] env[61868]: DEBUG nova.objects.instance [None req-da0b0370-9019-46e0-b203-afacf721d587 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lazy-loading 'resources' on Instance uuid a96cab5a-c053-4e45-96f6-2aba0a819110 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1066.787493] env[61868]: INFO nova.scheduler.client.report [None req-eedd44b4-3d1a-456b-a399-aa2e568c2d72 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Deleted allocations for instance 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1 [ 1067.298934] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eedd44b4-3d1a-456b-a399-aa2e568c2d72 tempest-ImagesTestJSON-1022884385 tempest-ImagesTestJSON-1022884385-project-member] Lock "7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.524s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.336023] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9da4d8bf-3ade-4068-a362-d3a2fd8708b1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.343333] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd8a285d-e8de-4a93-90d5-9a59db093f4f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.375183] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-776ea1ea-b266-4262-8133-8d68de9542fe {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.387999] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3069fe3-a0d4-419e-b111-aaddeb84d733 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.409879] env[61868]: DEBUG nova.compute.provider_tree [None req-da0b0370-9019-46e0-b203-afacf721d587 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1067.913098] env[61868]: DEBUG nova.scheduler.client.report [None req-da0b0370-9019-46e0-b203-afacf721d587 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1068.419882] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da0b0370-9019-46e0-b203-afacf721d587 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.653s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.447932] env[61868]: INFO nova.scheduler.client.report [None req-da0b0370-9019-46e0-b203-afacf721d587 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Deleted allocations for instance a96cab5a-c053-4e45-96f6-2aba0a819110 [ 1068.956666] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da0b0370-9019-46e0-b203-afacf721d587 tempest-AttachInterfacesTestJSON-133999971 tempest-AttachInterfacesTestJSON-133999971-project-member] Lock "a96cab5a-c053-4e45-96f6-2aba0a819110" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.441s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.597554] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "59171bf1-f2d4-4ff5-a661-025f5dca7600" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.598640] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "59171bf1-f2d4-4ff5-a661-025f5dca7600" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.104121] env[61868]: DEBUG nova.compute.manager [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1073.627031] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.627325] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.628865] env[61868]: INFO nova.compute.claims [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1073.635405] env[61868]: DEBUG oslo_concurrency.lockutils [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "f2d864f1-cd11-4b62-857d-789cf045f22c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.635665] env[61868]: DEBUG oslo_concurrency.lockutils [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "f2d864f1-cd11-4b62-857d-789cf045f22c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1074.138043] env[61868]: DEBUG nova.compute.manager [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1074.657280] env[61868]: DEBUG oslo_concurrency.lockutils [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.685764] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8f6ec4c-0201-4fe9-8dc9-4ae543949c90 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.694622] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-355cb5e7-e1d8-49e8-8413-9aef718c7fd1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.723444] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d87a6cd0-7b91-4e71-bc97-2a5e5501c947 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.730073] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f33f32a-2b6f-480d-a97d-f968c5587c61 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.742461] env[61868]: DEBUG nova.compute.provider_tree [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1075.246403] env[61868]: DEBUG nova.scheduler.client.report [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1075.751713] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.124s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.752234] env[61868]: DEBUG nova.compute.manager [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1075.755376] env[61868]: DEBUG oslo_concurrency.lockutils [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.098s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.756832] env[61868]: INFO nova.compute.claims [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1076.261925] env[61868]: DEBUG nova.compute.utils [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1076.264905] env[61868]: DEBUG nova.compute.manager [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1076.265090] env[61868]: DEBUG nova.network.neutron [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1076.304485] env[61868]: DEBUG nova.policy [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '11a64f7a464b45c192edd10486c61138', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bdab2203834c4ee5bab11aca70e48cc1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 1076.570696] env[61868]: DEBUG nova.network.neutron [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Successfully created port: 8f9ad20f-f367-4826-8aae-00a1192d0bc8 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1076.765811] env[61868]: DEBUG nova.compute.manager [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1076.825552] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78422812-d15d-494b-a70a-15ceb55aae99 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.833105] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aca1c029-e00c-4e4c-9d0b-127fb0edaf43 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.862656] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b19c535-7b9e-4827-aced-90bc2dbe9c05 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.869910] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3cc36c7-5875-4bb4-b0f8-da215cc99567 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.883066] env[61868]: DEBUG nova.compute.provider_tree [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1077.385849] env[61868]: DEBUG nova.scheduler.client.report [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1077.778912] env[61868]: DEBUG nova.compute.manager [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1077.803368] env[61868]: DEBUG nova.virt.hardware [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1077.803712] env[61868]: DEBUG nova.virt.hardware [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1077.803887] env[61868]: DEBUG nova.virt.hardware [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1077.804456] env[61868]: DEBUG nova.virt.hardware [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1077.804663] env[61868]: DEBUG nova.virt.hardware [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1077.804823] env[61868]: DEBUG nova.virt.hardware [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1077.805054] env[61868]: DEBUG nova.virt.hardware [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1077.805233] env[61868]: DEBUG nova.virt.hardware [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1077.805612] env[61868]: DEBUG nova.virt.hardware [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1077.805843] env[61868]: DEBUG nova.virt.hardware [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1077.806055] env[61868]: DEBUG nova.virt.hardware [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1077.806903] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-568ed4bc-f6f1-4bbb-9b49-1d78a1970b27 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.815264] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2041d138-ad51-4429-97b6-9295d65e8ae2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.890616] env[61868]: DEBUG oslo_concurrency.lockutils [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.135s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.891394] env[61868]: DEBUG nova.compute.manager [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1077.935014] env[61868]: DEBUG nova.compute.manager [req-a287acd2-4ece-4793-825b-c72aeaae3816 req-273f658e-a27f-4a5c-b133-5197c5726939 service nova] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Received event network-vif-plugged-8f9ad20f-f367-4826-8aae-00a1192d0bc8 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1077.935316] env[61868]: DEBUG oslo_concurrency.lockutils [req-a287acd2-4ece-4793-825b-c72aeaae3816 req-273f658e-a27f-4a5c-b133-5197c5726939 service nova] Acquiring lock "59171bf1-f2d4-4ff5-a661-025f5dca7600-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.935509] env[61868]: DEBUG oslo_concurrency.lockutils [req-a287acd2-4ece-4793-825b-c72aeaae3816 req-273f658e-a27f-4a5c-b133-5197c5726939 service nova] Lock "59171bf1-f2d4-4ff5-a661-025f5dca7600-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.935770] env[61868]: DEBUG oslo_concurrency.lockutils [req-a287acd2-4ece-4793-825b-c72aeaae3816 req-273f658e-a27f-4a5c-b133-5197c5726939 service nova] Lock "59171bf1-f2d4-4ff5-a661-025f5dca7600-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.935908] env[61868]: DEBUG nova.compute.manager [req-a287acd2-4ece-4793-825b-c72aeaae3816 req-273f658e-a27f-4a5c-b133-5197c5726939 service nova] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] No waiting events found dispatching network-vif-plugged-8f9ad20f-f367-4826-8aae-00a1192d0bc8 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1077.936076] env[61868]: WARNING nova.compute.manager [req-a287acd2-4ece-4793-825b-c72aeaae3816 req-273f658e-a27f-4a5c-b133-5197c5726939 service nova] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Received unexpected event network-vif-plugged-8f9ad20f-f367-4826-8aae-00a1192d0bc8 for instance with vm_state building and task_state spawning. [ 1078.025074] env[61868]: DEBUG nova.network.neutron [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Successfully updated port: 8f9ad20f-f367-4826-8aae-00a1192d0bc8 {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1078.397251] env[61868]: DEBUG nova.compute.utils [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1078.398697] env[61868]: DEBUG nova.compute.manager [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1078.398864] env[61868]: DEBUG nova.network.neutron [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1078.438478] env[61868]: DEBUG nova.policy [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4a981d6292f34cd7aba2108a9b9f3a90', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '28a7014141b24101ba7920bc83be4a9e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 1078.527814] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "refresh_cache-59171bf1-f2d4-4ff5-a661-025f5dca7600" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1078.527981] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquired lock "refresh_cache-59171bf1-f2d4-4ff5-a661-025f5dca7600" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1078.528160] env[61868]: DEBUG nova.network.neutron [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1078.686306] env[61868]: DEBUG nova.network.neutron [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Successfully created port: 4c07a7f4-8d47-494b-977a-1aa62477db98 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1078.903087] env[61868]: DEBUG nova.compute.manager [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1079.061935] env[61868]: DEBUG nova.network.neutron [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1079.184797] env[61868]: DEBUG nova.network.neutron [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Updating instance_info_cache with network_info: [{"id": "8f9ad20f-f367-4826-8aae-00a1192d0bc8", "address": "fa:16:3e:63:2a:68", "network": {"id": "16b3638c-1129-49b8-9f22-636ff7db7386", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1544179589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bdab2203834c4ee5bab11aca70e48cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d298db54-f13d-4bf6-b6c2-755074b3047f", "external-id": "nsx-vlan-transportzone-631", "segmentation_id": 631, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f9ad20f-f3", "ovs_interfaceid": "8f9ad20f-f367-4826-8aae-00a1192d0bc8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1079.686924] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Releasing lock "refresh_cache-59171bf1-f2d4-4ff5-a661-025f5dca7600" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1079.687266] env[61868]: DEBUG nova.compute.manager [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Instance network_info: |[{"id": "8f9ad20f-f367-4826-8aae-00a1192d0bc8", "address": "fa:16:3e:63:2a:68", "network": {"id": "16b3638c-1129-49b8-9f22-636ff7db7386", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1544179589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bdab2203834c4ee5bab11aca70e48cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d298db54-f13d-4bf6-b6c2-755074b3047f", "external-id": "nsx-vlan-transportzone-631", "segmentation_id": 631, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f9ad20f-f3", "ovs_interfaceid": "8f9ad20f-f367-4826-8aae-00a1192d0bc8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1079.687737] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:63:2a:68', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd298db54-f13d-4bf6-b6c2-755074b3047f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8f9ad20f-f367-4826-8aae-00a1192d0bc8', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1079.695477] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Creating folder: Project (bdab2203834c4ee5bab11aca70e48cc1). Parent ref: group-v281478. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1079.695767] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ed93fb1b-b178-43bb-9d96-d209f357b029 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.709243] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Created folder: Project (bdab2203834c4ee5bab11aca70e48cc1) in parent group-v281478. [ 1079.709438] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Creating folder: Instances. Parent ref: group-v281621. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1079.709659] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-601020fd-8180-4ad2-84c1-f522ec90ea63 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.718320] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Created folder: Instances in parent group-v281621. [ 1079.718577] env[61868]: DEBUG oslo.service.loopingcall [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1079.718805] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1079.718960] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b6a28d5d-b8a1-4655-81b9-93563c73c201 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.738013] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1079.738013] env[61868]: value = "task-1315985" [ 1079.738013] env[61868]: _type = "Task" [ 1079.738013] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.749169] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315985, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.913012] env[61868]: DEBUG nova.compute.manager [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1079.940802] env[61868]: DEBUG nova.virt.hardware [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1079.941102] env[61868]: DEBUG nova.virt.hardware [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1079.941285] env[61868]: DEBUG nova.virt.hardware [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1079.941493] env[61868]: DEBUG nova.virt.hardware [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1079.941659] env[61868]: DEBUG nova.virt.hardware [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1079.941815] env[61868]: DEBUG nova.virt.hardware [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1079.942048] env[61868]: DEBUG nova.virt.hardware [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1079.942235] env[61868]: DEBUG nova.virt.hardware [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1079.942410] env[61868]: DEBUG nova.virt.hardware [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1079.942577] env[61868]: DEBUG nova.virt.hardware [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1079.942755] env[61868]: DEBUG nova.virt.hardware [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1079.943675] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3ee460a-5ee6-402e-ad78-2af84459a8c6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.951689] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-386c5ec8-4314-413e-b4db-7cf1805497b3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.968649] env[61868]: DEBUG nova.compute.manager [req-5963d79e-22e5-48d7-889b-25fd7f8e5ea7 req-1c82f794-4a71-493e-82a9-348f720129f7 service nova] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Received event network-changed-8f9ad20f-f367-4826-8aae-00a1192d0bc8 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1079.968855] env[61868]: DEBUG nova.compute.manager [req-5963d79e-22e5-48d7-889b-25fd7f8e5ea7 req-1c82f794-4a71-493e-82a9-348f720129f7 service nova] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Refreshing instance network info cache due to event network-changed-8f9ad20f-f367-4826-8aae-00a1192d0bc8. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1079.969100] env[61868]: DEBUG oslo_concurrency.lockutils [req-5963d79e-22e5-48d7-889b-25fd7f8e5ea7 req-1c82f794-4a71-493e-82a9-348f720129f7 service nova] Acquiring lock "refresh_cache-59171bf1-f2d4-4ff5-a661-025f5dca7600" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1079.969263] env[61868]: DEBUG oslo_concurrency.lockutils [req-5963d79e-22e5-48d7-889b-25fd7f8e5ea7 req-1c82f794-4a71-493e-82a9-348f720129f7 service nova] Acquired lock "refresh_cache-59171bf1-f2d4-4ff5-a661-025f5dca7600" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1079.969430] env[61868]: DEBUG nova.network.neutron [req-5963d79e-22e5-48d7-889b-25fd7f8e5ea7 req-1c82f794-4a71-493e-82a9-348f720129f7 service nova] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Refreshing network info cache for port 8f9ad20f-f367-4826-8aae-00a1192d0bc8 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1080.249807] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315985, 'name': CreateVM_Task, 'duration_secs': 0.39141} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.250009] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1080.250850] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1080.251038] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1080.251437] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1080.251727] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b2638b60-c4ce-4d8a-b94d-c2d2ee207e18 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.256295] env[61868]: DEBUG oslo_vmware.api [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1080.256295] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52af00ff-db3e-5c2a-aa33-dc9db0899173" [ 1080.256295] env[61868]: _type = "Task" [ 1080.256295] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.263840] env[61868]: DEBUG oslo_vmware.api [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52af00ff-db3e-5c2a-aa33-dc9db0899173, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.320208] env[61868]: DEBUG nova.network.neutron [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Successfully updated port: 4c07a7f4-8d47-494b-977a-1aa62477db98 {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1080.664613] env[61868]: DEBUG nova.network.neutron [req-5963d79e-22e5-48d7-889b-25fd7f8e5ea7 req-1c82f794-4a71-493e-82a9-348f720129f7 service nova] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Updated VIF entry in instance network info cache for port 8f9ad20f-f367-4826-8aae-00a1192d0bc8. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1080.665013] env[61868]: DEBUG nova.network.neutron [req-5963d79e-22e5-48d7-889b-25fd7f8e5ea7 req-1c82f794-4a71-493e-82a9-348f720129f7 service nova] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Updating instance_info_cache with network_info: [{"id": "8f9ad20f-f367-4826-8aae-00a1192d0bc8", "address": "fa:16:3e:63:2a:68", "network": {"id": "16b3638c-1129-49b8-9f22-636ff7db7386", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1544179589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bdab2203834c4ee5bab11aca70e48cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d298db54-f13d-4bf6-b6c2-755074b3047f", "external-id": "nsx-vlan-transportzone-631", "segmentation_id": 631, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f9ad20f-f3", "ovs_interfaceid": "8f9ad20f-f367-4826-8aae-00a1192d0bc8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1080.767082] env[61868]: DEBUG oslo_vmware.api [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52af00ff-db3e-5c2a-aa33-dc9db0899173, 'name': SearchDatastore_Task, 'duration_secs': 0.009222} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.767082] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1080.767082] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1080.767284] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1080.767420] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1080.767516] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1080.767770] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-51a83bf6-b39c-448c-a3c7-ecc1fc7f3419 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.775706] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1080.775888] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1080.776563] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-357fa386-9271-4822-8b1f-a6b8507e13b1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.781390] env[61868]: DEBUG oslo_vmware.api [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1080.781390] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]528f5d39-e5dc-7e70-bed8-e75def920a81" [ 1080.781390] env[61868]: _type = "Task" [ 1080.781390] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.788349] env[61868]: DEBUG oslo_vmware.api [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]528f5d39-e5dc-7e70-bed8-e75def920a81, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.826415] env[61868]: DEBUG oslo_concurrency.lockutils [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "refresh_cache-f2d864f1-cd11-4b62-857d-789cf045f22c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1080.826554] env[61868]: DEBUG oslo_concurrency.lockutils [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquired lock "refresh_cache-f2d864f1-cd11-4b62-857d-789cf045f22c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1080.826717] env[61868]: DEBUG nova.network.neutron [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1081.167942] env[61868]: DEBUG oslo_concurrency.lockutils [req-5963d79e-22e5-48d7-889b-25fd7f8e5ea7 req-1c82f794-4a71-493e-82a9-348f720129f7 service nova] Releasing lock "refresh_cache-59171bf1-f2d4-4ff5-a661-025f5dca7600" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1081.292210] env[61868]: DEBUG oslo_vmware.api [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]528f5d39-e5dc-7e70-bed8-e75def920a81, 'name': SearchDatastore_Task, 'duration_secs': 0.009524} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.292992] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-516dce98-e16c-44fd-b4dd-4e7deac65c14 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.297927] env[61868]: DEBUG oslo_vmware.api [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1081.297927] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]528a4b06-0b7c-2465-7306-5901fecac1c7" [ 1081.297927] env[61868]: _type = "Task" [ 1081.297927] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.304985] env[61868]: DEBUG oslo_vmware.api [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]528a4b06-0b7c-2465-7306-5901fecac1c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.356743] env[61868]: DEBUG nova.network.neutron [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1081.476884] env[61868]: DEBUG nova.network.neutron [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Updating instance_info_cache with network_info: [{"id": "4c07a7f4-8d47-494b-977a-1aa62477db98", "address": "fa:16:3e:b7:7b:cc", "network": {"id": "7e8cbecf-dfdd-43f2-8745-ae1d2d55d309", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-33935587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a7014141b24101ba7920bc83be4a9e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4db2961d-273d-4634-9d06-a94fa9d384fb", "external-id": "nsx-vlan-transportzone-572", "segmentation_id": 572, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c07a7f4-8d", "ovs_interfaceid": "4c07a7f4-8d47-494b-977a-1aa62477db98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1081.809068] env[61868]: DEBUG oslo_vmware.api [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]528a4b06-0b7c-2465-7306-5901fecac1c7, 'name': SearchDatastore_Task, 'duration_secs': 0.010577} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.809068] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1081.809068] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 59171bf1-f2d4-4ff5-a661-025f5dca7600/59171bf1-f2d4-4ff5-a661-025f5dca7600.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1081.809339] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4443f10b-4d3c-4f0c-9a5c-b57c2095167f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.815782] env[61868]: DEBUG oslo_vmware.api [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1081.815782] env[61868]: value = "task-1315986" [ 1081.815782] env[61868]: _type = "Task" [ 1081.815782] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.823263] env[61868]: DEBUG oslo_vmware.api [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1315986, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.979643] env[61868]: DEBUG oslo_concurrency.lockutils [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Releasing lock "refresh_cache-f2d864f1-cd11-4b62-857d-789cf045f22c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1081.979978] env[61868]: DEBUG nova.compute.manager [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Instance network_info: |[{"id": "4c07a7f4-8d47-494b-977a-1aa62477db98", "address": "fa:16:3e:b7:7b:cc", "network": {"id": "7e8cbecf-dfdd-43f2-8745-ae1d2d55d309", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-33935587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a7014141b24101ba7920bc83be4a9e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4db2961d-273d-4634-9d06-a94fa9d384fb", "external-id": "nsx-vlan-transportzone-572", "segmentation_id": 572, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c07a7f4-8d", "ovs_interfaceid": "4c07a7f4-8d47-494b-977a-1aa62477db98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1081.980436] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b7:7b:cc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4db2961d-273d-4634-9d06-a94fa9d384fb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4c07a7f4-8d47-494b-977a-1aa62477db98', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1081.987940] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Creating folder: Project (28a7014141b24101ba7920bc83be4a9e). Parent ref: group-v281478. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1081.988279] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6d8566fb-ce23-4e11-9088-b1509018ceec {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.992587] env[61868]: DEBUG nova.compute.manager [req-45762d59-d606-42a0-a3a7-9f90dc7cdbdd req-787616fa-5b61-4f6d-9e6d-166653babe06 service nova] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Received event network-vif-plugged-4c07a7f4-8d47-494b-977a-1aa62477db98 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1081.992813] env[61868]: DEBUG oslo_concurrency.lockutils [req-45762d59-d606-42a0-a3a7-9f90dc7cdbdd req-787616fa-5b61-4f6d-9e6d-166653babe06 service nova] Acquiring lock "f2d864f1-cd11-4b62-857d-789cf045f22c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.993025] env[61868]: DEBUG oslo_concurrency.lockutils [req-45762d59-d606-42a0-a3a7-9f90dc7cdbdd req-787616fa-5b61-4f6d-9e6d-166653babe06 service nova] Lock "f2d864f1-cd11-4b62-857d-789cf045f22c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.993209] env[61868]: DEBUG oslo_concurrency.lockutils [req-45762d59-d606-42a0-a3a7-9f90dc7cdbdd req-787616fa-5b61-4f6d-9e6d-166653babe06 service nova] Lock "f2d864f1-cd11-4b62-857d-789cf045f22c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.993386] env[61868]: DEBUG nova.compute.manager [req-45762d59-d606-42a0-a3a7-9f90dc7cdbdd req-787616fa-5b61-4f6d-9e6d-166653babe06 service nova] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] No waiting events found dispatching network-vif-plugged-4c07a7f4-8d47-494b-977a-1aa62477db98 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1081.993585] env[61868]: WARNING nova.compute.manager [req-45762d59-d606-42a0-a3a7-9f90dc7cdbdd req-787616fa-5b61-4f6d-9e6d-166653babe06 service nova] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Received unexpected event network-vif-plugged-4c07a7f4-8d47-494b-977a-1aa62477db98 for instance with vm_state building and task_state spawning. [ 1081.994743] env[61868]: DEBUG nova.compute.manager [req-45762d59-d606-42a0-a3a7-9f90dc7cdbdd req-787616fa-5b61-4f6d-9e6d-166653babe06 service nova] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Received event network-changed-4c07a7f4-8d47-494b-977a-1aa62477db98 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1081.994743] env[61868]: DEBUG nova.compute.manager [req-45762d59-d606-42a0-a3a7-9f90dc7cdbdd req-787616fa-5b61-4f6d-9e6d-166653babe06 service nova] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Refreshing instance network info cache due to event network-changed-4c07a7f4-8d47-494b-977a-1aa62477db98. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1081.994743] env[61868]: DEBUG oslo_concurrency.lockutils [req-45762d59-d606-42a0-a3a7-9f90dc7cdbdd req-787616fa-5b61-4f6d-9e6d-166653babe06 service nova] Acquiring lock "refresh_cache-f2d864f1-cd11-4b62-857d-789cf045f22c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1081.994743] env[61868]: DEBUG oslo_concurrency.lockutils [req-45762d59-d606-42a0-a3a7-9f90dc7cdbdd req-787616fa-5b61-4f6d-9e6d-166653babe06 service nova] Acquired lock "refresh_cache-f2d864f1-cd11-4b62-857d-789cf045f22c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.994743] env[61868]: DEBUG nova.network.neutron [req-45762d59-d606-42a0-a3a7-9f90dc7cdbdd req-787616fa-5b61-4f6d-9e6d-166653babe06 service nova] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Refreshing network info cache for port 4c07a7f4-8d47-494b-977a-1aa62477db98 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1082.000429] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Created folder: Project (28a7014141b24101ba7920bc83be4a9e) in parent group-v281478. [ 1082.000646] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Creating folder: Instances. Parent ref: group-v281624. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1082.000905] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-38223609-80aa-4329-b0d0-4523ec87810f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.015425] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Created folder: Instances in parent group-v281624. [ 1082.015677] env[61868]: DEBUG oslo.service.loopingcall [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1082.015874] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1082.016113] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-efb1b186-ca07-4c63-9e6e-3fa3f06df72f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.035731] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1082.035731] env[61868]: value = "task-1315989" [ 1082.035731] env[61868]: _type = "Task" [ 1082.035731] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.044727] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315989, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.325524] env[61868]: DEBUG oslo_vmware.api [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1315986, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.483542} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.325870] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 59171bf1-f2d4-4ff5-a661-025f5dca7600/59171bf1-f2d4-4ff5-a661-025f5dca7600.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1082.326086] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1082.326354] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5e9effce-90ac-43c7-814e-05bc14ca2c0c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.331999] env[61868]: DEBUG oslo_vmware.api [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1082.331999] env[61868]: value = "task-1315990" [ 1082.331999] env[61868]: _type = "Task" [ 1082.331999] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.339765] env[61868]: DEBUG oslo_vmware.api [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1315990, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.545057] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1315989, 'name': CreateVM_Task, 'duration_secs': 0.435218} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.545217] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1082.545877] env[61868]: DEBUG oslo_concurrency.lockutils [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1082.546062] env[61868]: DEBUG oslo_concurrency.lockutils [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.546394] env[61868]: DEBUG oslo_concurrency.lockutils [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1082.546658] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-220c329c-29e7-4190-847a-d69a5771528d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.550645] env[61868]: DEBUG oslo_vmware.api [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1082.550645] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52c8801b-edf8-df87-1f35-32403aab18a3" [ 1082.550645] env[61868]: _type = "Task" [ 1082.550645] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.557651] env[61868]: DEBUG oslo_vmware.api [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52c8801b-edf8-df87-1f35-32403aab18a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.676556] env[61868]: DEBUG nova.network.neutron [req-45762d59-d606-42a0-a3a7-9f90dc7cdbdd req-787616fa-5b61-4f6d-9e6d-166653babe06 service nova] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Updated VIF entry in instance network info cache for port 4c07a7f4-8d47-494b-977a-1aa62477db98. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1082.676951] env[61868]: DEBUG nova.network.neutron [req-45762d59-d606-42a0-a3a7-9f90dc7cdbdd req-787616fa-5b61-4f6d-9e6d-166653babe06 service nova] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Updating instance_info_cache with network_info: [{"id": "4c07a7f4-8d47-494b-977a-1aa62477db98", "address": "fa:16:3e:b7:7b:cc", "network": {"id": "7e8cbecf-dfdd-43f2-8745-ae1d2d55d309", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-33935587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a7014141b24101ba7920bc83be4a9e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4db2961d-273d-4634-9d06-a94fa9d384fb", "external-id": "nsx-vlan-transportzone-572", "segmentation_id": 572, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c07a7f4-8d", "ovs_interfaceid": "4c07a7f4-8d47-494b-977a-1aa62477db98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1082.841517] env[61868]: DEBUG oslo_vmware.api [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1315990, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.098454} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.841759] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1082.842539] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f23c701-ebb1-4c26-8263-ac34b7cb7bc5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.863268] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] 59171bf1-f2d4-4ff5-a661-025f5dca7600/59171bf1-f2d4-4ff5-a661-025f5dca7600.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1082.863532] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8e05abda-1cb0-490a-972e-741325c4e68f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.881474] env[61868]: DEBUG oslo_vmware.api [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1082.881474] env[61868]: value = "task-1315991" [ 1082.881474] env[61868]: _type = "Task" [ 1082.881474] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.888575] env[61868]: DEBUG oslo_vmware.api [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1315991, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.060269] env[61868]: DEBUG oslo_vmware.api [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52c8801b-edf8-df87-1f35-32403aab18a3, 'name': SearchDatastore_Task, 'duration_secs': 0.008542} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.060637] env[61868]: DEBUG oslo_concurrency.lockutils [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.060831] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1083.061082] env[61868]: DEBUG oslo_concurrency.lockutils [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.061241] env[61868]: DEBUG oslo_concurrency.lockutils [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.061451] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1083.061710] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4b3c6a22-be26-4bb6-bd7d-3c80aeeaad63 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.070058] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1083.070246] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1083.070917] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-057f1d13-b13c-45c2-af12-67510bfad1ff {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.075547] env[61868]: DEBUG oslo_vmware.api [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1083.075547] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]523922e8-273c-5508-12aa-f8e2b8f72d72" [ 1083.075547] env[61868]: _type = "Task" [ 1083.075547] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.084192] env[61868]: DEBUG oslo_vmware.api [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]523922e8-273c-5508-12aa-f8e2b8f72d72, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.179311] env[61868]: DEBUG oslo_concurrency.lockutils [req-45762d59-d606-42a0-a3a7-9f90dc7cdbdd req-787616fa-5b61-4f6d-9e6d-166653babe06 service nova] Releasing lock "refresh_cache-f2d864f1-cd11-4b62-857d-789cf045f22c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.390930] env[61868]: DEBUG oslo_vmware.api [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1315991, 'name': ReconfigVM_Task, 'duration_secs': 0.302727} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.391271] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Reconfigured VM instance instance-00000067 to attach disk [datastore2] 59171bf1-f2d4-4ff5-a661-025f5dca7600/59171bf1-f2d4-4ff5-a661-025f5dca7600.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1083.391875] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6b071303-fac3-4dbc-b8a4-7d1328ace9cf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.397200] env[61868]: DEBUG oslo_vmware.api [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1083.397200] env[61868]: value = "task-1315992" [ 1083.397200] env[61868]: _type = "Task" [ 1083.397200] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.404617] env[61868]: DEBUG oslo_vmware.api [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1315992, 'name': Rename_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.585311] env[61868]: DEBUG oslo_vmware.api [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]523922e8-273c-5508-12aa-f8e2b8f72d72, 'name': SearchDatastore_Task, 'duration_secs': 0.008127} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.586038] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-28112765-5fef-46ad-84c0-c116c47d1fcc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.591237] env[61868]: DEBUG oslo_vmware.api [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1083.591237] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]523d825a-25b2-514f-64c8-b6ce9bbb82e9" [ 1083.591237] env[61868]: _type = "Task" [ 1083.591237] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.598301] env[61868]: DEBUG oslo_vmware.api [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]523d825a-25b2-514f-64c8-b6ce9bbb82e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.907210] env[61868]: DEBUG oslo_vmware.api [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1315992, 'name': Rename_Task, 'duration_secs': 0.129614} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.907560] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1083.907825] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8cd4d690-8704-4e53-a1f8-c1c374c7be1e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.914386] env[61868]: DEBUG oslo_vmware.api [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1083.914386] env[61868]: value = "task-1315993" [ 1083.914386] env[61868]: _type = "Task" [ 1083.914386] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.921735] env[61868]: DEBUG oslo_vmware.api [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1315993, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.101515] env[61868]: DEBUG oslo_vmware.api [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]523d825a-25b2-514f-64c8-b6ce9bbb82e9, 'name': SearchDatastore_Task, 'duration_secs': 0.011104} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.101751] env[61868]: DEBUG oslo_concurrency.lockutils [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.102059] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] f2d864f1-cd11-4b62-857d-789cf045f22c/f2d864f1-cd11-4b62-857d-789cf045f22c.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1084.102335] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-99c4b289-2336-4176-b443-ba92579612e0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.109055] env[61868]: DEBUG oslo_vmware.api [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1084.109055] env[61868]: value = "task-1315994" [ 1084.109055] env[61868]: _type = "Task" [ 1084.109055] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.116854] env[61868]: DEBUG oslo_vmware.api [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1315994, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.427538] env[61868]: DEBUG oslo_vmware.api [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1315993, 'name': PowerOnVM_Task, 'duration_secs': 0.453906} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.427888] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1084.428058] env[61868]: INFO nova.compute.manager [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Took 6.65 seconds to spawn the instance on the hypervisor. [ 1084.428253] env[61868]: DEBUG nova.compute.manager [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1084.429071] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cb267f7-2567-4eae-8999-ea5c569324e6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.620466] env[61868]: DEBUG oslo_vmware.api [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1315994, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.494366} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.620753] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] f2d864f1-cd11-4b62-857d-789cf045f22c/f2d864f1-cd11-4b62-857d-789cf045f22c.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1084.620937] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1084.621206] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-01741b9b-5ce5-4f85-bb3d-19512eca7301 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.627409] env[61868]: DEBUG oslo_vmware.api [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1084.627409] env[61868]: value = "task-1315995" [ 1084.627409] env[61868]: _type = "Task" [ 1084.627409] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.635470] env[61868]: DEBUG oslo_vmware.api [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1315995, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.950520] env[61868]: INFO nova.compute.manager [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Took 11.34 seconds to build instance. [ 1085.137363] env[61868]: DEBUG oslo_vmware.api [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1315995, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067831} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.138824] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1085.138824] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dde18e71-6126-4588-8eaf-ac8873b4bc04 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.161620] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] f2d864f1-cd11-4b62-857d-789cf045f22c/f2d864f1-cd11-4b62-857d-789cf045f22c.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1085.161870] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a8acfe63-659b-49ed-b012-98a129bd07a7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.180721] env[61868]: DEBUG oslo_vmware.api [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1085.180721] env[61868]: value = "task-1315996" [ 1085.180721] env[61868]: _type = "Task" [ 1085.180721] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.188225] env[61868]: DEBUG oslo_vmware.api [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1315996, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.453213] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c3fc95e9-70b3-4567-84ea-44452a6e220b tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "59171bf1-f2d4-4ff5-a661-025f5dca7600" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.854s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.501775] env[61868]: DEBUG nova.compute.manager [req-f558076e-b0bd-4d35-8244-0da532784b7b req-77eeb42b-34dd-46b0-a636-1b10074fc9da service nova] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Received event network-changed-8f9ad20f-f367-4826-8aae-00a1192d0bc8 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1085.501989] env[61868]: DEBUG nova.compute.manager [req-f558076e-b0bd-4d35-8244-0da532784b7b req-77eeb42b-34dd-46b0-a636-1b10074fc9da service nova] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Refreshing instance network info cache due to event network-changed-8f9ad20f-f367-4826-8aae-00a1192d0bc8. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1085.502220] env[61868]: DEBUG oslo_concurrency.lockutils [req-f558076e-b0bd-4d35-8244-0da532784b7b req-77eeb42b-34dd-46b0-a636-1b10074fc9da service nova] Acquiring lock "refresh_cache-59171bf1-f2d4-4ff5-a661-025f5dca7600" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1085.502360] env[61868]: DEBUG oslo_concurrency.lockutils [req-f558076e-b0bd-4d35-8244-0da532784b7b req-77eeb42b-34dd-46b0-a636-1b10074fc9da service nova] Acquired lock "refresh_cache-59171bf1-f2d4-4ff5-a661-025f5dca7600" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1085.502708] env[61868]: DEBUG nova.network.neutron [req-f558076e-b0bd-4d35-8244-0da532784b7b req-77eeb42b-34dd-46b0-a636-1b10074fc9da service nova] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Refreshing network info cache for port 8f9ad20f-f367-4826-8aae-00a1192d0bc8 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1085.690879] env[61868]: DEBUG oslo_vmware.api [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1315996, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.191615] env[61868]: DEBUG oslo_vmware.api [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1315996, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.229813] env[61868]: DEBUG nova.network.neutron [req-f558076e-b0bd-4d35-8244-0da532784b7b req-77eeb42b-34dd-46b0-a636-1b10074fc9da service nova] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Updated VIF entry in instance network info cache for port 8f9ad20f-f367-4826-8aae-00a1192d0bc8. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1086.230269] env[61868]: DEBUG nova.network.neutron [req-f558076e-b0bd-4d35-8244-0da532784b7b req-77eeb42b-34dd-46b0-a636-1b10074fc9da service nova] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Updating instance_info_cache with network_info: [{"id": "8f9ad20f-f367-4826-8aae-00a1192d0bc8", "address": "fa:16:3e:63:2a:68", "network": {"id": "16b3638c-1129-49b8-9f22-636ff7db7386", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1544179589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.190", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bdab2203834c4ee5bab11aca70e48cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d298db54-f13d-4bf6-b6c2-755074b3047f", "external-id": "nsx-vlan-transportzone-631", "segmentation_id": 631, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f9ad20f-f3", "ovs_interfaceid": "8f9ad20f-f367-4826-8aae-00a1192d0bc8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1086.691632] env[61868]: DEBUG oslo_vmware.api [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1315996, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.733409] env[61868]: DEBUG oslo_concurrency.lockutils [req-f558076e-b0bd-4d35-8244-0da532784b7b req-77eeb42b-34dd-46b0-a636-1b10074fc9da service nova] Releasing lock "refresh_cache-59171bf1-f2d4-4ff5-a661-025f5dca7600" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1087.192839] env[61868]: DEBUG oslo_vmware.api [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1315996, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.693558] env[61868]: DEBUG oslo_vmware.api [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1315996, 'name': ReconfigVM_Task, 'duration_secs': 2.09692} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.693957] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Reconfigured VM instance instance-00000068 to attach disk [datastore2] f2d864f1-cd11-4b62-857d-789cf045f22c/f2d864f1-cd11-4b62-857d-789cf045f22c.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1087.694493] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f6796585-9e2b-4ee8-a568-7e5e50989660 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.700706] env[61868]: DEBUG oslo_vmware.api [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1087.700706] env[61868]: value = "task-1315997" [ 1087.700706] env[61868]: _type = "Task" [ 1087.700706] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.708181] env[61868]: DEBUG oslo_vmware.api [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1315997, 'name': Rename_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.210911] env[61868]: DEBUG oslo_vmware.api [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1315997, 'name': Rename_Task, 'duration_secs': 0.398059} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.211194] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1088.211436] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-18fcfb04-8de2-469a-9777-2e46df6a357e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.217935] env[61868]: DEBUG oslo_vmware.api [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1088.217935] env[61868]: value = "task-1315998" [ 1088.217935] env[61868]: _type = "Task" [ 1088.217935] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.225831] env[61868]: DEBUG oslo_vmware.api [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1315998, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.728501] env[61868]: DEBUG oslo_vmware.api [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1315998, 'name': PowerOnVM_Task, 'duration_secs': 0.477124} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.728894] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1088.729091] env[61868]: INFO nova.compute.manager [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Took 8.82 seconds to spawn the instance on the hypervisor. [ 1088.729307] env[61868]: DEBUG nova.compute.manager [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1088.730050] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dead0532-1665-4962-bddc-179e96657144 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.247926] env[61868]: INFO nova.compute.manager [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Took 14.61 seconds to build instance. [ 1089.749768] env[61868]: DEBUG oslo_concurrency.lockutils [None req-dfd66713-9260-4a0b-89e2-771947428413 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "f2d864f1-cd11-4b62-857d-789cf045f22c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.114s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.036471] env[61868]: DEBUG nova.compute.manager [req-3c386556-0e93-4983-b891-a7a0c2bcb066 req-92b4e282-768a-44a1-b999-e0c32a14c9ad service nova] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Received event network-changed-4c07a7f4-8d47-494b-977a-1aa62477db98 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1090.036696] env[61868]: DEBUG nova.compute.manager [req-3c386556-0e93-4983-b891-a7a0c2bcb066 req-92b4e282-768a-44a1-b999-e0c32a14c9ad service nova] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Refreshing instance network info cache due to event network-changed-4c07a7f4-8d47-494b-977a-1aa62477db98. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1090.036944] env[61868]: DEBUG oslo_concurrency.lockutils [req-3c386556-0e93-4983-b891-a7a0c2bcb066 req-92b4e282-768a-44a1-b999-e0c32a14c9ad service nova] Acquiring lock "refresh_cache-f2d864f1-cd11-4b62-857d-789cf045f22c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1090.037123] env[61868]: DEBUG oslo_concurrency.lockutils [req-3c386556-0e93-4983-b891-a7a0c2bcb066 req-92b4e282-768a-44a1-b999-e0c32a14c9ad service nova] Acquired lock "refresh_cache-f2d864f1-cd11-4b62-857d-789cf045f22c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1090.037302] env[61868]: DEBUG nova.network.neutron [req-3c386556-0e93-4983-b891-a7a0c2bcb066 req-92b4e282-768a-44a1-b999-e0c32a14c9ad service nova] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Refreshing network info cache for port 4c07a7f4-8d47-494b-977a-1aa62477db98 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1090.737319] env[61868]: DEBUG nova.network.neutron [req-3c386556-0e93-4983-b891-a7a0c2bcb066 req-92b4e282-768a-44a1-b999-e0c32a14c9ad service nova] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Updated VIF entry in instance network info cache for port 4c07a7f4-8d47-494b-977a-1aa62477db98. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1090.737713] env[61868]: DEBUG nova.network.neutron [req-3c386556-0e93-4983-b891-a7a0c2bcb066 req-92b4e282-768a-44a1-b999-e0c32a14c9ad service nova] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Updating instance_info_cache with network_info: [{"id": "4c07a7f4-8d47-494b-977a-1aa62477db98", "address": "fa:16:3e:b7:7b:cc", "network": {"id": "7e8cbecf-dfdd-43f2-8745-ae1d2d55d309", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-33935587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a7014141b24101ba7920bc83be4a9e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4db2961d-273d-4634-9d06-a94fa9d384fb", "external-id": "nsx-vlan-transportzone-572", "segmentation_id": 572, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c07a7f4-8d", "ovs_interfaceid": "4c07a7f4-8d47-494b-977a-1aa62477db98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1091.240355] env[61868]: DEBUG oslo_concurrency.lockutils [req-3c386556-0e93-4983-b891-a7a0c2bcb066 req-92b4e282-768a-44a1-b999-e0c32a14c9ad service nova] Releasing lock "refresh_cache-f2d864f1-cd11-4b62-857d-789cf045f22c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.075055] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1095.075475] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1095.075475] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1095.075558] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1095.075697] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61868) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1095.075841] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager.update_available_resource {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1095.579185] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.579418] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.579590] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.579755] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61868) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1095.580723] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04803084-cc38-4287-a9f4-4988096a3601 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.588976] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-971ffd9d-60b6-4943-b6f9-e68bd3715ab2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.602496] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e65acd0-9769-4d84-bb3a-40ec8df460dc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.608547] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d86df75e-ac9b-424a-8f2c-47a41178a2fe {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.638344] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180164MB free_disk=134GB free_vcpus=48 pci_devices=None {{(pid=61868) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1095.638496] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.638693] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.664817] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1096.665072] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 59171bf1-f2d4-4ff5-a661-025f5dca7600 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1096.665108] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance f2d864f1-cd11-4b62-857d-789cf045f22c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1096.665286] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=61868) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1096.665423] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=61868) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1096.707924] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebb37a45-0a55-4ab6-b8e7-14bd9450d609 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.715130] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7047dad9-20a5-48bb-aa99-51cc74551e91 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.743780] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ff34cd4-2abc-4f0d-b07d-6812af8c047f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.750563] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd8e39bd-7bdd-4c88-b68d-8bdf340d0636 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.762758] env[61868]: DEBUG nova.compute.provider_tree [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1097.265858] env[61868]: DEBUG nova.scheduler.client.report [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1097.770467] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61868) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1097.770900] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.132s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.770900] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.771028] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Cleaning up deleted instances {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1098.286325] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] There are 49 instances to clean {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1098.286609] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 1cb04c98-f725-4b1f-933e-ded580d340fb] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1098.790218] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 3a94eb16-cce3-4765-bc61-b6776bdd5b63] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1099.293542] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 7ff48de2-e5b3-4c49-b8b1-dd0d2f697ac1] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1099.797304] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 0316bfec-128e-41dc-ad3a-9ba3844b95fa] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1100.301263] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: d9e90641-6765-4bda-ab07-3506898f06a9] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1100.805096] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: c96c4bcb-03f9-4961-aa17-8364aa4cf66e] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1101.308853] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 69a0b94b-2bf4-49ad-8fb0-ea945176e5eb] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1101.659511] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3ddcff15-ace1-406d-bc15-edecde5a781f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "273a38e7-8eb0-46fc-b9e2-31a9a7fc34db" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.659783] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3ddcff15-ace1-406d-bc15-edecde5a781f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "273a38e7-8eb0-46fc-b9e2-31a9a7fc34db" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.812756] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 54dacd73-c760-49bd-9baf-ad2012a53264] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1102.162843] env[61868]: DEBUG nova.compute.utils [None req-3ddcff15-ace1-406d-bc15-edecde5a781f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1102.315624] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 56703761-3421-44fd-aa63-9db36c86e69c] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1102.667043] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3ddcff15-ace1-406d-bc15-edecde5a781f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "273a38e7-8eb0-46fc-b9e2-31a9a7fc34db" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.819147] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: eeeec752-7e99-471d-9425-c4ccf4f6d0a5] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1103.322519] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: cbeeb7b1-166b-49f7-88c5-6ed05c9dc759] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1103.721297] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3ddcff15-ace1-406d-bc15-edecde5a781f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "273a38e7-8eb0-46fc-b9e2-31a9a7fc34db" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1103.721543] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3ddcff15-ace1-406d-bc15-edecde5a781f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "273a38e7-8eb0-46fc-b9e2-31a9a7fc34db" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1103.722056] env[61868]: INFO nova.compute.manager [None req-3ddcff15-ace1-406d-bc15-edecde5a781f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Attaching volume b1bf590f-f7ba-4bd0-b207-2e0c9996fc62 to /dev/sdb [ 1103.753679] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-320bcff8-7eff-4a55-8058-35b4bb13c76c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.761277] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd6a6431-db5b-4ec3-b355-c1a5faf289c0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.774257] env[61868]: DEBUG nova.virt.block_device [None req-3ddcff15-ace1-406d-bc15-edecde5a781f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Updating existing volume attachment record: 59c3c944-b624-4e57-9347-2b981aa69472 {{(pid=61868) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1103.825723] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: a96cab5a-c053-4e45-96f6-2aba0a819110] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1104.329256] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 9286cc76-1152-4fd4-9279-33d6ee4b9486] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1104.833307] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 5d2b56d5-a4e0-4dba-9158-71036d60a40b] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1105.336589] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 4216e9b6-7b86-409e-9543-1c57e0f4b850] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1105.839862] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: cc2b4281-fd3a-4db5-ba33-bfbe873a1e9b] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1106.343473] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 867384c6-175b-4848-b9ac-917470742ba2] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1106.846567] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 87b25494-1853-4c7b-ae8b-6283992e895c] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1107.349597] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 6ce39c07-939c-49f3-8871-6e2163709f4a] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1107.853079] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 4f70e3f8-3a36-4d01-b3fc-73b9428fd2e9] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1108.316974] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ddcff15-ace1-406d-bc15-edecde5a781f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Volume attach. Driver type: vmdk {{(pid=61868) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1108.317260] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ddcff15-ace1-406d-bc15-edecde5a781f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281627', 'volume_id': 'b1bf590f-f7ba-4bd0-b207-2e0c9996fc62', 'name': 'volume-b1bf590f-f7ba-4bd0-b207-2e0c9996fc62', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '273a38e7-8eb0-46fc-b9e2-31a9a7fc34db', 'attached_at': '', 'detached_at': '', 'volume_id': 'b1bf590f-f7ba-4bd0-b207-2e0c9996fc62', 'serial': 'b1bf590f-f7ba-4bd0-b207-2e0c9996fc62'} {{(pid=61868) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1108.318163] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3a68b59-01e2-40a4-be21-ac4c3a5e307a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.335474] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffd68f3b-a611-4d2c-be13-a403220571d8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.360090] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ddcff15-ace1-406d-bc15-edecde5a781f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] volume-b1bf590f-f7ba-4bd0-b207-2e0c9996fc62/volume-b1bf590f-f7ba-4bd0-b207-2e0c9996fc62.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1108.360560] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 4e9d75e8-ca33-4e33-baa3-cb7d596f584e] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1108.362340] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c3bca2f3-dcc0-4ee4-8a74-c5c6993caad9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.380924] env[61868]: DEBUG oslo_vmware.api [None req-3ddcff15-ace1-406d-bc15-edecde5a781f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 1108.380924] env[61868]: value = "task-1316001" [ 1108.380924] env[61868]: _type = "Task" [ 1108.380924] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.390374] env[61868]: DEBUG oslo_vmware.api [None req-3ddcff15-ace1-406d-bc15-edecde5a781f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316001, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.876421] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: b51e968e-6911-47f3-b5a9-e8a5865e42a0] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1108.889834] env[61868]: DEBUG oslo_vmware.api [None req-3ddcff15-ace1-406d-bc15-edecde5a781f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316001, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.380629] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: e59a182d-97b2-454a-bc40-8afb0839324f] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1109.391471] env[61868]: DEBUG oslo_vmware.api [None req-3ddcff15-ace1-406d-bc15-edecde5a781f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316001, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.887073] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 842c8606-632a-47d4-9a67-38d8e6ce6b4d] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1109.894258] env[61868]: DEBUG oslo_vmware.api [None req-3ddcff15-ace1-406d-bc15-edecde5a781f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316001, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.392640] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 8ea8e28b-3582-45f2-b4b5-84f624415a58] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1110.394745] env[61868]: DEBUG oslo_vmware.api [None req-3ddcff15-ace1-406d-bc15-edecde5a781f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316001, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.893078] env[61868]: DEBUG oslo_vmware.api [None req-3ddcff15-ace1-406d-bc15-edecde5a781f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316001, 'name': ReconfigVM_Task, 'duration_secs': 2.081229} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.893450] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ddcff15-ace1-406d-bc15-edecde5a781f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Reconfigured VM instance instance-00000066 to attach disk [datastore2] volume-b1bf590f-f7ba-4bd0-b207-2e0c9996fc62/volume-b1bf590f-f7ba-4bd0-b207-2e0c9996fc62.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1110.898248] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: e2832c0f-dc62-4628-87d4-e2a01819b771] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1110.899896] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-37c4803b-171d-49b6-9a8a-823f66fc6315 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.915052] env[61868]: DEBUG oslo_vmware.api [None req-3ddcff15-ace1-406d-bc15-edecde5a781f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 1110.915052] env[61868]: value = "task-1316002" [ 1110.915052] env[61868]: _type = "Task" [ 1110.915052] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.923657] env[61868]: DEBUG oslo_vmware.api [None req-3ddcff15-ace1-406d-bc15-edecde5a781f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316002, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.410724] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 306f2dc4-933c-4948-8330-eda93931c4e0] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1111.424617] env[61868]: DEBUG oslo_vmware.api [None req-3ddcff15-ace1-406d-bc15-edecde5a781f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316002, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.914682] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: f1c5ba09-d739-4e3f-9edf-3fab3f2f39c3] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1111.925673] env[61868]: DEBUG oslo_vmware.api [None req-3ddcff15-ace1-406d-bc15-edecde5a781f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316002, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.421779] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: b7055e5a-37d2-42d5-bab0-1e70faf52b58] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1112.428971] env[61868]: DEBUG oslo_vmware.api [None req-3ddcff15-ace1-406d-bc15-edecde5a781f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316002, 'name': ReconfigVM_Task, 'duration_secs': 1.131605} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.429253] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ddcff15-ace1-406d-bc15-edecde5a781f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281627', 'volume_id': 'b1bf590f-f7ba-4bd0-b207-2e0c9996fc62', 'name': 'volume-b1bf590f-f7ba-4bd0-b207-2e0c9996fc62', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '273a38e7-8eb0-46fc-b9e2-31a9a7fc34db', 'attached_at': '', 'detached_at': '', 'volume_id': 'b1bf590f-f7ba-4bd0-b207-2e0c9996fc62', 'serial': 'b1bf590f-f7ba-4bd0-b207-2e0c9996fc62'} {{(pid=61868) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1112.925471] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: b9b5be37-6b30-4229-9c8d-3ee9d30db119] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1113.429171] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: c6108cf2-b597-4ca7-8f57-12ea82cc6c2c] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1113.462806] env[61868]: DEBUG nova.objects.instance [None req-3ddcff15-ace1-406d-bc15-edecde5a781f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lazy-loading 'flavor' on Instance uuid 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1113.932217] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 5aec2609-79d3-4725-a182-70b575adbe17] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1113.968323] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3ddcff15-ace1-406d-bc15-edecde5a781f tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "273a38e7-8eb0-46fc-b9e2-31a9a7fc34db" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 10.247s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.252152] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3b7b6f69-025d-4169-bbf7-f816b8daa897 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "273a38e7-8eb0-46fc-b9e2-31a9a7fc34db" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.252387] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3b7b6f69-025d-4169-bbf7-f816b8daa897 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "273a38e7-8eb0-46fc-b9e2-31a9a7fc34db" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.435576] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: cf063dbc-7b5a-4836-91bf-a0aa33cca6bc] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1114.755343] env[61868]: INFO nova.compute.manager [None req-3b7b6f69-025d-4169-bbf7-f816b8daa897 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Detaching volume b1bf590f-f7ba-4bd0-b207-2e0c9996fc62 [ 1114.785829] env[61868]: INFO nova.virt.block_device [None req-3b7b6f69-025d-4169-bbf7-f816b8daa897 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Attempting to driver detach volume b1bf590f-f7ba-4bd0-b207-2e0c9996fc62 from mountpoint /dev/sdb [ 1114.786125] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b7b6f69-025d-4169-bbf7-f816b8daa897 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Volume detach. Driver type: vmdk {{(pid=61868) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1114.786336] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b7b6f69-025d-4169-bbf7-f816b8daa897 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281627', 'volume_id': 'b1bf590f-f7ba-4bd0-b207-2e0c9996fc62', 'name': 'volume-b1bf590f-f7ba-4bd0-b207-2e0c9996fc62', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '273a38e7-8eb0-46fc-b9e2-31a9a7fc34db', 'attached_at': '', 'detached_at': '', 'volume_id': 'b1bf590f-f7ba-4bd0-b207-2e0c9996fc62', 'serial': 'b1bf590f-f7ba-4bd0-b207-2e0c9996fc62'} {{(pid=61868) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1114.787230] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9be9a527-9c60-4628-81f3-4aefc612b75f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.808397] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcb22843-1491-45ff-b973-a8d0a8694874 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.815020] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acec242e-1152-4705-9bc0-1c5f6ab2858a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.835020] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35de317c-0802-4da4-a8e4-fcaf9c5b9dd1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.849951] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b7b6f69-025d-4169-bbf7-f816b8daa897 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] The volume has not been displaced from its original location: [datastore2] volume-b1bf590f-f7ba-4bd0-b207-2e0c9996fc62/volume-b1bf590f-f7ba-4bd0-b207-2e0c9996fc62.vmdk. No consolidation needed. {{(pid=61868) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1114.855328] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b7b6f69-025d-4169-bbf7-f816b8daa897 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Reconfiguring VM instance instance-00000066 to detach disk 2001 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1114.855612] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3ca0efe9-0c1a-4aeb-b38a-c84d096bdd40 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.873284] env[61868]: DEBUG oslo_vmware.api [None req-3b7b6f69-025d-4169-bbf7-f816b8daa897 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 1114.873284] env[61868]: value = "task-1316003" [ 1114.873284] env[61868]: _type = "Task" [ 1114.873284] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.880617] env[61868]: DEBUG oslo_vmware.api [None req-3b7b6f69-025d-4169-bbf7-f816b8daa897 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316003, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.938518] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: c6f99cbc-3e81-44b4-b3d1-d8cd821829f2] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1115.385063] env[61868]: DEBUG oslo_vmware.api [None req-3b7b6f69-025d-4169-bbf7-f816b8daa897 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316003, 'name': ReconfigVM_Task, 'duration_secs': 0.20115} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.385349] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b7b6f69-025d-4169-bbf7-f816b8daa897 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Reconfigured VM instance instance-00000066 to detach disk 2001 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1115.390084] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cb41d0ef-264b-4abf-87b3-c5e5229c4a90 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.404905] env[61868]: DEBUG oslo_vmware.api [None req-3b7b6f69-025d-4169-bbf7-f816b8daa897 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 1115.404905] env[61868]: value = "task-1316004" [ 1115.404905] env[61868]: _type = "Task" [ 1115.404905] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.415077] env[61868]: DEBUG oslo_vmware.api [None req-3b7b6f69-025d-4169-bbf7-f816b8daa897 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316004, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.441742] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 75fdbc11-1ef3-442d-bcf4-55069ba9ab62] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1115.914423] env[61868]: DEBUG oslo_vmware.api [None req-3b7b6f69-025d-4169-bbf7-f816b8daa897 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316004, 'name': ReconfigVM_Task, 'duration_secs': 0.127017} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.914709] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b7b6f69-025d-4169-bbf7-f816b8daa897 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281627', 'volume_id': 'b1bf590f-f7ba-4bd0-b207-2e0c9996fc62', 'name': 'volume-b1bf590f-f7ba-4bd0-b207-2e0c9996fc62', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '273a38e7-8eb0-46fc-b9e2-31a9a7fc34db', 'attached_at': '', 'detached_at': '', 'volume_id': 'b1bf590f-f7ba-4bd0-b207-2e0c9996fc62', 'serial': 'b1bf590f-f7ba-4bd0-b207-2e0c9996fc62'} {{(pid=61868) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1115.944691] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 19c2720c-90bc-47f6-999b-6031f893408d] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1116.448258] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: ca85dfd1-e794-41dc-a337-fbd99998e1fa] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1116.454177] env[61868]: DEBUG nova.objects.instance [None req-3b7b6f69-025d-4169-bbf7-f816b8daa897 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lazy-loading 'flavor' on Instance uuid 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1116.951611] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: c7b10943-5136-44c8-b4e7-59651220a333] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1117.455515] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 409cad1a-946d-4c58-aa57-1c0bf97fe63f] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1117.460248] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3b7b6f69-025d-4169-bbf7-f816b8daa897 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "273a38e7-8eb0-46fc-b9e2-31a9a7fc34db" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.208s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.958737] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: c2693a43-4ea2-4ab0-8915-2fa544780e3c] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1118.462242] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 7788f9f8-c08e-46f8-b204-070215223436] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1118.489302] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "273a38e7-8eb0-46fc-b9e2-31a9a7fc34db" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1118.489542] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "273a38e7-8eb0-46fc-b9e2-31a9a7fc34db" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.489744] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "273a38e7-8eb0-46fc-b9e2-31a9a7fc34db-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1118.489932] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "273a38e7-8eb0-46fc-b9e2-31a9a7fc34db-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.490118] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "273a38e7-8eb0-46fc-b9e2-31a9a7fc34db-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.491942] env[61868]: INFO nova.compute.manager [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Terminating instance [ 1118.493612] env[61868]: DEBUG nova.compute.manager [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1118.493806] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1118.494681] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea75740e-37c5-49f0-9d4f-2eb19c7c37bb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.502418] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1118.502640] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c93ae0a4-6051-45a7-874f-fd37a3162302 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.508523] env[61868]: DEBUG oslo_vmware.api [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 1118.508523] env[61868]: value = "task-1316005" [ 1118.508523] env[61868]: _type = "Task" [ 1118.508523] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.515988] env[61868]: DEBUG oslo_vmware.api [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316005, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.965852] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 52a5afc6-4e79-436c-bc94-b61ca9fb860c] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1119.019056] env[61868]: DEBUG oslo_vmware.api [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316005, 'name': PowerOffVM_Task, 'duration_secs': 0.200497} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.019327] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1119.019514] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1119.019989] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-307a9f0f-4734-4645-a31b-c72a92dc5c76 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.079476] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1119.079709] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1119.079911] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Deleting the datastore file [datastore1] 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1119.080187] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d2a14824-5d31-431b-8676-ca2c279df942 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.086642] env[61868]: DEBUG oslo_vmware.api [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 1119.086642] env[61868]: value = "task-1316007" [ 1119.086642] env[61868]: _type = "Task" [ 1119.086642] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.093855] env[61868]: DEBUG oslo_vmware.api [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316007, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.469554] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 3dff312c-b7d2-4b1a-ab9b-0dc03bbfd1f1] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1119.596316] env[61868]: DEBUG oslo_vmware.api [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316007, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.121948} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.596605] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1119.596812] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1119.597011] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1119.597226] env[61868]: INFO nova.compute.manager [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1119.597553] env[61868]: DEBUG oslo.service.loopingcall [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1119.597770] env[61868]: DEBUG nova.compute.manager [-] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1119.597868] env[61868]: DEBUG nova.network.neutron [-] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1119.972495] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 0a7cbce9-0a9b-4421-8bab-d2c4bd976fbc] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1120.068872] env[61868]: DEBUG nova.compute.manager [req-5872aa10-2fbf-43d9-8993-ea9243f6f069 req-606ad083-23f2-4c7d-8785-993516dec5b7 service nova] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Received event network-vif-deleted-a34b4784-533a-42f9-b5c9-ea8332e8eecb {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1120.069138] env[61868]: INFO nova.compute.manager [req-5872aa10-2fbf-43d9-8993-ea9243f6f069 req-606ad083-23f2-4c7d-8785-993516dec5b7 service nova] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Neutron deleted interface a34b4784-533a-42f9-b5c9-ea8332e8eecb; detaching it from the instance and deleting it from the info cache [ 1120.069416] env[61868]: DEBUG nova.network.neutron [req-5872aa10-2fbf-43d9-8993-ea9243f6f069 req-606ad083-23f2-4c7d-8785-993516dec5b7 service nova] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1120.476101] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 02cd5bd0-1bf2-4f34-9fb7-c21a7a76bd02] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1120.545490] env[61868]: DEBUG nova.network.neutron [-] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1120.572219] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f88c91dc-7a24-4abd-a0b3-45f59508786d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.581926] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b1263cd-90c4-4153-ac46-d50d8a48f490 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.604957] env[61868]: DEBUG nova.compute.manager [req-5872aa10-2fbf-43d9-8993-ea9243f6f069 req-606ad083-23f2-4c7d-8785-993516dec5b7 service nova] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Detach interface failed, port_id=a34b4784-533a-42f9-b5c9-ea8332e8eecb, reason: Instance 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db could not be found. {{(pid=61868) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1120.979600] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: d35953d8-692d-498e-baf5-96ef381ce12b] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1121.048084] env[61868]: INFO nova.compute.manager [-] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Took 1.45 seconds to deallocate network for instance. [ 1121.483111] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 2c00210e-3b6a-4041-9e8e-7f40ef23d716] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1121.554269] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1121.554541] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1121.554764] env[61868]: DEBUG nova.objects.instance [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lazy-loading 'resources' on Instance uuid 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1121.986745] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 083140bf-233d-49ac-8247-cd93206f9200] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1122.166899] env[61868]: DEBUG nova.scheduler.client.report [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Refreshing inventories for resource provider 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1122.180607] env[61868]: DEBUG nova.scheduler.client.report [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Updating ProviderTree inventory for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1122.180829] env[61868]: DEBUG nova.compute.provider_tree [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Updating inventory in ProviderTree for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1122.190605] env[61868]: DEBUG nova.scheduler.client.report [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Refreshing aggregate associations for resource provider 83e67721-2ac9-4a23-aa31-82aca86979c8, aggregates: None {{(pid=61868) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1122.206794] env[61868]: DEBUG nova.scheduler.client.report [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Refreshing trait associations for resource provider 83e67721-2ac9-4a23-aa31-82aca86979c8, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61868) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1122.250648] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1d85412-411c-4128-8a1d-f261075ff948 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.257790] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d985f121-4371-402a-9407-8b6b6d3583e1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.287945] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b792e8e-927b-4834-8e8e-b76733d07316 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.294588] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f15aaa8b-2b47-4834-b1d6-0d3fab4357a2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.307251] env[61868]: DEBUG nova.compute.provider_tree [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1122.490426] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 4de32ee6-e205-435b-a75e-94c3121048e3] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1122.810484] env[61868]: DEBUG nova.scheduler.client.report [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1122.993663] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1122.993923] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Cleaning up deleted instances with incomplete migration {{(pid=61868) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1123.316115] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.761s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.335174] env[61868]: INFO nova.scheduler.client.report [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Deleted allocations for instance 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db [ 1123.496614] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1123.627072] env[61868]: DEBUG oslo_concurrency.lockutils [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "04b0079b-8f5c-4e51-8bb9-52d81273713e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.627320] env[61868]: DEBUG oslo_concurrency.lockutils [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "04b0079b-8f5c-4e51-8bb9-52d81273713e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.842314] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3c9bc11e-c1fb-4e89-bc41-54b409ab6904 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "273a38e7-8eb0-46fc-b9e2-31a9a7fc34db" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.353s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.130146] env[61868]: DEBUG nova.compute.manager [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1124.655929] env[61868]: DEBUG oslo_concurrency.lockutils [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.656234] env[61868]: DEBUG oslo_concurrency.lockutils [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1124.657837] env[61868]: INFO nova.compute.claims [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1124.998068] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1125.314052] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "afd7ff53-f164-4f3d-9ece-28a10245e38c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1125.314449] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "afd7ff53-f164-4f3d-9ece-28a10245e38c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1125.505273] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1125.505273] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Starting heal instance info cache {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1125.720439] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c90ec8e8-572e-4891-8fde-3a1b3eeda4ad {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.727816] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1d6346d-b3d8-49db-8266-3ff44816f408 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.757381] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a45dce65-612b-46a3-a862-7a8848550739 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.764483] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-400cbc99-3b46-4f30-a59f-04bad8c5b0e9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.777566] env[61868]: DEBUG nova.compute.provider_tree [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1125.816158] env[61868]: DEBUG nova.compute.manager [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1126.280802] env[61868]: DEBUG nova.scheduler.client.report [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1126.333919] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1126.509609] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Didn't find any instances for network info cache update. {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1126.509837] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1126.509998] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1126.510182] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1126.785107] env[61868]: DEBUG oslo_concurrency.lockutils [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.129s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.785727] env[61868]: DEBUG nova.compute.manager [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1126.788312] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.454s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1126.789754] env[61868]: INFO nova.compute.claims [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1127.293582] env[61868]: DEBUG nova.compute.utils [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1127.298259] env[61868]: DEBUG nova.compute.manager [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1127.298259] env[61868]: DEBUG nova.network.neutron [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1127.336974] env[61868]: DEBUG nova.policy [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '11a64f7a464b45c192edd10486c61138', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bdab2203834c4ee5bab11aca70e48cc1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 1127.576475] env[61868]: DEBUG nova.network.neutron [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Successfully created port: bf8801b5-7f14-408d-95bd-deab0e42905a {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1127.720329] env[61868]: DEBUG nova.compute.manager [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1127.721301] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84c32169-0cb1-427b-8b2e-f095e6ce2436 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.799019] env[61868]: DEBUG nova.compute.manager [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1127.863482] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-056af72d-175b-46ad-ac8a-a60b9f43e245 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.872553] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-874823e5-5db9-4782-9f93-ddb0dad95022 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.902976] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccdc2a94-ba3d-489b-918b-f8357c3836e5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.909111] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d10f5cf-fc89-452f-bc38-cbf189b1250a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.922008] env[61868]: DEBUG nova.compute.provider_tree [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1128.232200] env[61868]: INFO nova.compute.manager [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] instance snapshotting [ 1128.232883] env[61868]: DEBUG nova.objects.instance [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lazy-loading 'flavor' on Instance uuid f2d864f1-cd11-4b62-857d-789cf045f22c {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1128.425615] env[61868]: DEBUG nova.scheduler.client.report [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1128.738352] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dc0d9ce-075a-47e3-996c-8a0b198bca84 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.758048] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95f65791-ca26-4c38-b7b5-3258ec34e7b9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.807545] env[61868]: DEBUG nova.compute.manager [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1128.828883] env[61868]: DEBUG nova.virt.hardware [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1128.829136] env[61868]: DEBUG nova.virt.hardware [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1128.829302] env[61868]: DEBUG nova.virt.hardware [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1128.829493] env[61868]: DEBUG nova.virt.hardware [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1128.829644] env[61868]: DEBUG nova.virt.hardware [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1128.829796] env[61868]: DEBUG nova.virt.hardware [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1128.830011] env[61868]: DEBUG nova.virt.hardware [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1128.830190] env[61868]: DEBUG nova.virt.hardware [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1128.830363] env[61868]: DEBUG nova.virt.hardware [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1128.830629] env[61868]: DEBUG nova.virt.hardware [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1128.830716] env[61868]: DEBUG nova.virt.hardware [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1128.831570] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9c4dc00-1b4e-4703-b635-39bfa3a67556 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.839732] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94b4dd21-f0a5-498c-9486-514a198bc2a9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.930459] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.142s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1128.931068] env[61868]: DEBUG nova.compute.manager [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1128.950796] env[61868]: DEBUG nova.compute.manager [req-a27a8999-5b3f-48eb-92de-18239cc01d66 req-188d782f-6da9-4474-bfb1-88b08a4a04dd service nova] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Received event network-vif-plugged-bf8801b5-7f14-408d-95bd-deab0e42905a {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1128.950796] env[61868]: DEBUG oslo_concurrency.lockutils [req-a27a8999-5b3f-48eb-92de-18239cc01d66 req-188d782f-6da9-4474-bfb1-88b08a4a04dd service nova] Acquiring lock "04b0079b-8f5c-4e51-8bb9-52d81273713e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1128.950796] env[61868]: DEBUG oslo_concurrency.lockutils [req-a27a8999-5b3f-48eb-92de-18239cc01d66 req-188d782f-6da9-4474-bfb1-88b08a4a04dd service nova] Lock "04b0079b-8f5c-4e51-8bb9-52d81273713e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1128.950796] env[61868]: DEBUG oslo_concurrency.lockutils [req-a27a8999-5b3f-48eb-92de-18239cc01d66 req-188d782f-6da9-4474-bfb1-88b08a4a04dd service nova] Lock "04b0079b-8f5c-4e51-8bb9-52d81273713e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1128.950796] env[61868]: DEBUG nova.compute.manager [req-a27a8999-5b3f-48eb-92de-18239cc01d66 req-188d782f-6da9-4474-bfb1-88b08a4a04dd service nova] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] No waiting events found dispatching network-vif-plugged-bf8801b5-7f14-408d-95bd-deab0e42905a {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1128.952146] env[61868]: WARNING nova.compute.manager [req-a27a8999-5b3f-48eb-92de-18239cc01d66 req-188d782f-6da9-4474-bfb1-88b08a4a04dd service nova] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Received unexpected event network-vif-plugged-bf8801b5-7f14-408d-95bd-deab0e42905a for instance with vm_state building and task_state spawning. [ 1129.035874] env[61868]: DEBUG nova.network.neutron [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Successfully updated port: bf8801b5-7f14-408d-95bd-deab0e42905a {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1129.267625] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Creating Snapshot of the VM instance {{(pid=61868) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1129.267940] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-2e0886b0-81ae-42cc-a603-a09fffe0ebe7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.276063] env[61868]: DEBUG oslo_vmware.api [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1129.276063] env[61868]: value = "task-1316008" [ 1129.276063] env[61868]: _type = "Task" [ 1129.276063] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.284048] env[61868]: DEBUG oslo_vmware.api [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316008, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.437348] env[61868]: DEBUG nova.compute.utils [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1129.438787] env[61868]: DEBUG nova.compute.manager [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1129.438950] env[61868]: DEBUG nova.network.neutron [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1129.477172] env[61868]: DEBUG nova.policy [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0fa71d6f8941441d89709e7e3af1ed9f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c8440ce785c344d793dd513b5f008d65', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 1129.538751] env[61868]: DEBUG oslo_concurrency.lockutils [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "refresh_cache-04b0079b-8f5c-4e51-8bb9-52d81273713e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1129.538920] env[61868]: DEBUG oslo_concurrency.lockutils [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquired lock "refresh_cache-04b0079b-8f5c-4e51-8bb9-52d81273713e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1129.538997] env[61868]: DEBUG nova.network.neutron [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1129.722210] env[61868]: DEBUG nova.network.neutron [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Successfully created port: 4f02e3ef-73cc-42e9-92a7-4366bb33e757 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1129.785723] env[61868]: DEBUG oslo_vmware.api [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316008, 'name': CreateSnapshot_Task, 'duration_secs': 0.441369} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.786010] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Created Snapshot of the VM instance {{(pid=61868) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1129.786893] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c784594-5d87-4f56-96e4-daa06a93d194 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.941673] env[61868]: DEBUG nova.compute.manager [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1130.082131] env[61868]: DEBUG nova.network.neutron [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1130.218459] env[61868]: DEBUG nova.network.neutron [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Updating instance_info_cache with network_info: [{"id": "bf8801b5-7f14-408d-95bd-deab0e42905a", "address": "fa:16:3e:e7:b3:ac", "network": {"id": "16b3638c-1129-49b8-9f22-636ff7db7386", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1544179589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bdab2203834c4ee5bab11aca70e48cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d298db54-f13d-4bf6-b6c2-755074b3047f", "external-id": "nsx-vlan-transportzone-631", "segmentation_id": 631, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf8801b5-7f", "ovs_interfaceid": "bf8801b5-7f14-408d-95bd-deab0e42905a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1130.304351] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Creating linked-clone VM from snapshot {{(pid=61868) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1130.304682] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-2a122559-ddb6-4d5d-808c-535bec0b4303 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.313394] env[61868]: DEBUG oslo_vmware.api [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1130.313394] env[61868]: value = "task-1316009" [ 1130.313394] env[61868]: _type = "Task" [ 1130.313394] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.322079] env[61868]: DEBUG oslo_vmware.api [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316009, 'name': CloneVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.721446] env[61868]: DEBUG oslo_concurrency.lockutils [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Releasing lock "refresh_cache-04b0079b-8f5c-4e51-8bb9-52d81273713e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1130.721891] env[61868]: DEBUG nova.compute.manager [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Instance network_info: |[{"id": "bf8801b5-7f14-408d-95bd-deab0e42905a", "address": "fa:16:3e:e7:b3:ac", "network": {"id": "16b3638c-1129-49b8-9f22-636ff7db7386", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1544179589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bdab2203834c4ee5bab11aca70e48cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d298db54-f13d-4bf6-b6c2-755074b3047f", "external-id": "nsx-vlan-transportzone-631", "segmentation_id": 631, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf8801b5-7f", "ovs_interfaceid": "bf8801b5-7f14-408d-95bd-deab0e42905a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1130.722574] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e7:b3:ac', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd298db54-f13d-4bf6-b6c2-755074b3047f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bf8801b5-7f14-408d-95bd-deab0e42905a', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1130.730464] env[61868]: DEBUG oslo.service.loopingcall [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1130.730710] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1130.731490] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a745bd5d-e8d4-4cc2-8fe5-871cc2575992 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.750637] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1130.750637] env[61868]: value = "task-1316010" [ 1130.750637] env[61868]: _type = "Task" [ 1130.750637] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.757873] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1316010, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.822681] env[61868]: DEBUG oslo_vmware.api [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316009, 'name': CloneVM_Task} progress is 94%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.952132] env[61868]: DEBUG nova.compute.manager [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1130.973649] env[61868]: DEBUG nova.virt.hardware [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1130.974388] env[61868]: DEBUG nova.virt.hardware [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1130.974388] env[61868]: DEBUG nova.virt.hardware [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1130.974388] env[61868]: DEBUG nova.virt.hardware [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1130.974587] env[61868]: DEBUG nova.virt.hardware [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1130.975100] env[61868]: DEBUG nova.virt.hardware [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1130.975100] env[61868]: DEBUG nova.virt.hardware [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1130.975100] env[61868]: DEBUG nova.virt.hardware [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1130.975273] env[61868]: DEBUG nova.virt.hardware [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1130.975428] env[61868]: DEBUG nova.virt.hardware [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1130.975613] env[61868]: DEBUG nova.virt.hardware [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1130.976775] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-132c8571-6df4-403a-8a80-ef7c42047ae4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.981215] env[61868]: DEBUG nova.compute.manager [req-577e906b-e39f-4c53-8678-2aea1582223b req-8cdc4d76-14ed-43ec-bb86-092583878241 service nova] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Received event network-changed-bf8801b5-7f14-408d-95bd-deab0e42905a {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1130.981406] env[61868]: DEBUG nova.compute.manager [req-577e906b-e39f-4c53-8678-2aea1582223b req-8cdc4d76-14ed-43ec-bb86-092583878241 service nova] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Refreshing instance network info cache due to event network-changed-bf8801b5-7f14-408d-95bd-deab0e42905a. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1130.981806] env[61868]: DEBUG oslo_concurrency.lockutils [req-577e906b-e39f-4c53-8678-2aea1582223b req-8cdc4d76-14ed-43ec-bb86-092583878241 service nova] Acquiring lock "refresh_cache-04b0079b-8f5c-4e51-8bb9-52d81273713e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1130.981806] env[61868]: DEBUG oslo_concurrency.lockutils [req-577e906b-e39f-4c53-8678-2aea1582223b req-8cdc4d76-14ed-43ec-bb86-092583878241 service nova] Acquired lock "refresh_cache-04b0079b-8f5c-4e51-8bb9-52d81273713e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1130.981932] env[61868]: DEBUG nova.network.neutron [req-577e906b-e39f-4c53-8678-2aea1582223b req-8cdc4d76-14ed-43ec-bb86-092583878241 service nova] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Refreshing network info cache for port bf8801b5-7f14-408d-95bd-deab0e42905a {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1130.989290] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21426c9e-9419-4a72-8563-af1622a11cd0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.200288] env[61868]: DEBUG nova.network.neutron [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Successfully updated port: 4f02e3ef-73cc-42e9-92a7-4366bb33e757 {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1131.260881] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1316010, 'name': CreateVM_Task, 'duration_secs': 0.441678} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.261058] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1131.261723] env[61868]: DEBUG oslo_concurrency.lockutils [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1131.261893] env[61868]: DEBUG oslo_concurrency.lockutils [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.262291] env[61868]: DEBUG oslo_concurrency.lockutils [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1131.262554] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12931b54-a4a5-4c87-b7e2-99c5a345ead0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.267111] env[61868]: DEBUG oslo_vmware.api [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1131.267111] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5228294b-1668-8e0c-a015-0b268d309b53" [ 1131.267111] env[61868]: _type = "Task" [ 1131.267111] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.275074] env[61868]: DEBUG oslo_vmware.api [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5228294b-1668-8e0c-a015-0b268d309b53, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.322793] env[61868]: DEBUG oslo_vmware.api [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316009, 'name': CloneVM_Task, 'duration_secs': 0.950557} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.324047] env[61868]: INFO nova.virt.vmwareapi.vmops [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Created linked-clone VM from snapshot [ 1131.324047] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8293d76-6bb1-46da-a1da-1cc2ef60207c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.330898] env[61868]: DEBUG nova.virt.vmwareapi.images [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Uploading image e70ee3c8-6ebd-4457-8ba6-ef94cd5a1e56 {{(pid=61868) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1131.350183] env[61868]: DEBUG oslo_vmware.rw_handles [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1131.350183] env[61868]: value = "vm-281629" [ 1131.350183] env[61868]: _type = "VirtualMachine" [ 1131.350183] env[61868]: }. {{(pid=61868) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1131.350420] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-7117560a-4448-4010-8da6-7cef69bb39f3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.356448] env[61868]: DEBUG oslo_vmware.rw_handles [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lease: (returnval){ [ 1131.356448] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]521e72b1-497d-58ad-29fd-0871c746cd7d" [ 1131.356448] env[61868]: _type = "HttpNfcLease" [ 1131.356448] env[61868]: } obtained for exporting VM: (result){ [ 1131.356448] env[61868]: value = "vm-281629" [ 1131.356448] env[61868]: _type = "VirtualMachine" [ 1131.356448] env[61868]: }. {{(pid=61868) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1131.356692] env[61868]: DEBUG oslo_vmware.api [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the lease: (returnval){ [ 1131.356692] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]521e72b1-497d-58ad-29fd-0871c746cd7d" [ 1131.356692] env[61868]: _type = "HttpNfcLease" [ 1131.356692] env[61868]: } to be ready. {{(pid=61868) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1131.364103] env[61868]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1131.364103] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]521e72b1-497d-58ad-29fd-0871c746cd7d" [ 1131.364103] env[61868]: _type = "HttpNfcLease" [ 1131.364103] env[61868]: } is initializing. {{(pid=61868) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1131.663415] env[61868]: DEBUG nova.network.neutron [req-577e906b-e39f-4c53-8678-2aea1582223b req-8cdc4d76-14ed-43ec-bb86-092583878241 service nova] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Updated VIF entry in instance network info cache for port bf8801b5-7f14-408d-95bd-deab0e42905a. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1131.663796] env[61868]: DEBUG nova.network.neutron [req-577e906b-e39f-4c53-8678-2aea1582223b req-8cdc4d76-14ed-43ec-bb86-092583878241 service nova] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Updating instance_info_cache with network_info: [{"id": "bf8801b5-7f14-408d-95bd-deab0e42905a", "address": "fa:16:3e:e7:b3:ac", "network": {"id": "16b3638c-1129-49b8-9f22-636ff7db7386", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1544179589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bdab2203834c4ee5bab11aca70e48cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d298db54-f13d-4bf6-b6c2-755074b3047f", "external-id": "nsx-vlan-transportzone-631", "segmentation_id": 631, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf8801b5-7f", "ovs_interfaceid": "bf8801b5-7f14-408d-95bd-deab0e42905a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1131.706530] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "refresh_cache-afd7ff53-f164-4f3d-9ece-28a10245e38c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1131.706795] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquired lock "refresh_cache-afd7ff53-f164-4f3d-9ece-28a10245e38c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.706957] env[61868]: DEBUG nova.network.neutron [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1131.777256] env[61868]: DEBUG oslo_vmware.api [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5228294b-1668-8e0c-a015-0b268d309b53, 'name': SearchDatastore_Task, 'duration_secs': 0.01013} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.777567] env[61868]: DEBUG oslo_concurrency.lockutils [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1131.777754] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1131.778009] env[61868]: DEBUG oslo_concurrency.lockutils [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1131.778171] env[61868]: DEBUG oslo_concurrency.lockutils [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.778357] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1131.778637] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-73b87748-9c36-499d-b073-8b3bfd4629fb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.787108] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1131.787286] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1131.787943] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d438737a-ab12-4c5a-9291-7a97e7f8803c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.792619] env[61868]: DEBUG oslo_vmware.api [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1131.792619] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52c97a68-ab5f-8bf2-587e-d99b1a91f2d8" [ 1131.792619] env[61868]: _type = "Task" [ 1131.792619] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.799744] env[61868]: DEBUG oslo_vmware.api [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52c97a68-ab5f-8bf2-587e-d99b1a91f2d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.863741] env[61868]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1131.863741] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]521e72b1-497d-58ad-29fd-0871c746cd7d" [ 1131.863741] env[61868]: _type = "HttpNfcLease" [ 1131.863741] env[61868]: } is ready. {{(pid=61868) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1131.864039] env[61868]: DEBUG oslo_vmware.rw_handles [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1131.864039] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]521e72b1-497d-58ad-29fd-0871c746cd7d" [ 1131.864039] env[61868]: _type = "HttpNfcLease" [ 1131.864039] env[61868]: }. {{(pid=61868) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1131.864837] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97a50e27-5090-4bd3-9406-4d76ed31a55c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.871855] env[61868]: DEBUG oslo_vmware.rw_handles [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52154e2c-a08a-7555-550d-8baa8c79b013/disk-0.vmdk from lease info. {{(pid=61868) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1131.872045] env[61868]: DEBUG oslo_vmware.rw_handles [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52154e2c-a08a-7555-550d-8baa8c79b013/disk-0.vmdk for reading. {{(pid=61868) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1131.972288] env[61868]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-9b20c16b-bb19-49de-a869-bba2bfa6f938 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.166696] env[61868]: DEBUG oslo_concurrency.lockutils [req-577e906b-e39f-4c53-8678-2aea1582223b req-8cdc4d76-14ed-43ec-bb86-092583878241 service nova] Releasing lock "refresh_cache-04b0079b-8f5c-4e51-8bb9-52d81273713e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1132.244363] env[61868]: DEBUG nova.network.neutron [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1132.302875] env[61868]: DEBUG oslo_vmware.api [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52c97a68-ab5f-8bf2-587e-d99b1a91f2d8, 'name': SearchDatastore_Task, 'duration_secs': 0.009548} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.303745] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9f4792f-bfd7-4014-849c-56c97e9a8412 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.312706] env[61868]: DEBUG oslo_vmware.api [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1132.312706] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52d82cb5-f3e4-3e2d-267f-1723030d0d43" [ 1132.312706] env[61868]: _type = "Task" [ 1132.312706] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.320565] env[61868]: DEBUG oslo_vmware.api [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52d82cb5-f3e4-3e2d-267f-1723030d0d43, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.385113] env[61868]: DEBUG nova.network.neutron [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Updating instance_info_cache with network_info: [{"id": "4f02e3ef-73cc-42e9-92a7-4366bb33e757", "address": "fa:16:3e:59:03:b9", "network": {"id": "3535e048-fe57-4cbc-9489-0b77484069bd", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-748518849-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c8440ce785c344d793dd513b5f008d65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f02e3ef-73", "ovs_interfaceid": "4f02e3ef-73cc-42e9-92a7-4366bb33e757", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1132.822943] env[61868]: DEBUG oslo_vmware.api [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52d82cb5-f3e4-3e2d-267f-1723030d0d43, 'name': SearchDatastore_Task, 'duration_secs': 0.010044} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.823316] env[61868]: DEBUG oslo_concurrency.lockutils [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1132.823644] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 04b0079b-8f5c-4e51-8bb9-52d81273713e/04b0079b-8f5c-4e51-8bb9-52d81273713e.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1132.823935] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-213d6e80-601a-4cad-93f8-7402617c2853 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.831341] env[61868]: DEBUG oslo_vmware.api [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1132.831341] env[61868]: value = "task-1316012" [ 1132.831341] env[61868]: _type = "Task" [ 1132.831341] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.838795] env[61868]: DEBUG oslo_vmware.api [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316012, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.887680] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Releasing lock "refresh_cache-afd7ff53-f164-4f3d-9ece-28a10245e38c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1132.887978] env[61868]: DEBUG nova.compute.manager [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Instance network_info: |[{"id": "4f02e3ef-73cc-42e9-92a7-4366bb33e757", "address": "fa:16:3e:59:03:b9", "network": {"id": "3535e048-fe57-4cbc-9489-0b77484069bd", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-748518849-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c8440ce785c344d793dd513b5f008d65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f02e3ef-73", "ovs_interfaceid": "4f02e3ef-73cc-42e9-92a7-4366bb33e757", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1132.888464] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:59:03:b9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'da0e5087-d65b-416f-90fe-beaa9c534ad3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4f02e3ef-73cc-42e9-92a7-4366bb33e757', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1132.896503] env[61868]: DEBUG oslo.service.loopingcall [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1132.896877] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1132.897237] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-25cbba74-fa0f-4277-9578-3070b2ea2aa3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.918196] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1132.918196] env[61868]: value = "task-1316013" [ 1132.918196] env[61868]: _type = "Task" [ 1132.918196] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.926185] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1316013, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.011220] env[61868]: DEBUG nova.compute.manager [req-6c41ae39-bfe5-43a5-b18b-060f52adf237 req-527a7fbf-2834-4297-9085-439d2c007e85 service nova] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Received event network-vif-plugged-4f02e3ef-73cc-42e9-92a7-4366bb33e757 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1133.011661] env[61868]: DEBUG oslo_concurrency.lockutils [req-6c41ae39-bfe5-43a5-b18b-060f52adf237 req-527a7fbf-2834-4297-9085-439d2c007e85 service nova] Acquiring lock "afd7ff53-f164-4f3d-9ece-28a10245e38c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1133.012120] env[61868]: DEBUG oslo_concurrency.lockutils [req-6c41ae39-bfe5-43a5-b18b-060f52adf237 req-527a7fbf-2834-4297-9085-439d2c007e85 service nova] Lock "afd7ff53-f164-4f3d-9ece-28a10245e38c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.012422] env[61868]: DEBUG oslo_concurrency.lockutils [req-6c41ae39-bfe5-43a5-b18b-060f52adf237 req-527a7fbf-2834-4297-9085-439d2c007e85 service nova] Lock "afd7ff53-f164-4f3d-9ece-28a10245e38c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.012764] env[61868]: DEBUG nova.compute.manager [req-6c41ae39-bfe5-43a5-b18b-060f52adf237 req-527a7fbf-2834-4297-9085-439d2c007e85 service nova] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] No waiting events found dispatching network-vif-plugged-4f02e3ef-73cc-42e9-92a7-4366bb33e757 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1133.013099] env[61868]: WARNING nova.compute.manager [req-6c41ae39-bfe5-43a5-b18b-060f52adf237 req-527a7fbf-2834-4297-9085-439d2c007e85 service nova] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Received unexpected event network-vif-plugged-4f02e3ef-73cc-42e9-92a7-4366bb33e757 for instance with vm_state building and task_state spawning. [ 1133.013424] env[61868]: DEBUG nova.compute.manager [req-6c41ae39-bfe5-43a5-b18b-060f52adf237 req-527a7fbf-2834-4297-9085-439d2c007e85 service nova] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Received event network-changed-4f02e3ef-73cc-42e9-92a7-4366bb33e757 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1133.013711] env[61868]: DEBUG nova.compute.manager [req-6c41ae39-bfe5-43a5-b18b-060f52adf237 req-527a7fbf-2834-4297-9085-439d2c007e85 service nova] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Refreshing instance network info cache due to event network-changed-4f02e3ef-73cc-42e9-92a7-4366bb33e757. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1133.013922] env[61868]: DEBUG oslo_concurrency.lockutils [req-6c41ae39-bfe5-43a5-b18b-060f52adf237 req-527a7fbf-2834-4297-9085-439d2c007e85 service nova] Acquiring lock "refresh_cache-afd7ff53-f164-4f3d-9ece-28a10245e38c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1133.014106] env[61868]: DEBUG oslo_concurrency.lockutils [req-6c41ae39-bfe5-43a5-b18b-060f52adf237 req-527a7fbf-2834-4297-9085-439d2c007e85 service nova] Acquired lock "refresh_cache-afd7ff53-f164-4f3d-9ece-28a10245e38c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1133.014491] env[61868]: DEBUG nova.network.neutron [req-6c41ae39-bfe5-43a5-b18b-060f52adf237 req-527a7fbf-2834-4297-9085-439d2c007e85 service nova] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Refreshing network info cache for port 4f02e3ef-73cc-42e9-92a7-4366bb33e757 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1133.342018] env[61868]: DEBUG oslo_vmware.api [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316012, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.474916} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.342388] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 04b0079b-8f5c-4e51-8bb9-52d81273713e/04b0079b-8f5c-4e51-8bb9-52d81273713e.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1133.342750] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1133.343105] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-643b3add-f2f1-49b0-bf99-5982da1e4bb3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.349187] env[61868]: DEBUG oslo_vmware.api [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1133.349187] env[61868]: value = "task-1316014" [ 1133.349187] env[61868]: _type = "Task" [ 1133.349187] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.357119] env[61868]: DEBUG oslo_vmware.api [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316014, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.428635] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1316013, 'name': CreateVM_Task, 'duration_secs': 0.454555} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.428867] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1133.429629] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1133.429841] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1133.430464] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1133.430748] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-acf3daa6-16a5-4f29-aae2-86ee4fd7b37b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.435524] env[61868]: DEBUG oslo_vmware.api [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 1133.435524] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5251cc3c-f98a-314b-5cd0-a1fe1e2a9928" [ 1133.435524] env[61868]: _type = "Task" [ 1133.435524] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.443306] env[61868]: DEBUG oslo_vmware.api [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5251cc3c-f98a-314b-5cd0-a1fe1e2a9928, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.716319] env[61868]: DEBUG nova.network.neutron [req-6c41ae39-bfe5-43a5-b18b-060f52adf237 req-527a7fbf-2834-4297-9085-439d2c007e85 service nova] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Updated VIF entry in instance network info cache for port 4f02e3ef-73cc-42e9-92a7-4366bb33e757. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1133.716705] env[61868]: DEBUG nova.network.neutron [req-6c41ae39-bfe5-43a5-b18b-060f52adf237 req-527a7fbf-2834-4297-9085-439d2c007e85 service nova] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Updating instance_info_cache with network_info: [{"id": "4f02e3ef-73cc-42e9-92a7-4366bb33e757", "address": "fa:16:3e:59:03:b9", "network": {"id": "3535e048-fe57-4cbc-9489-0b77484069bd", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-748518849-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c8440ce785c344d793dd513b5f008d65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f02e3ef-73", "ovs_interfaceid": "4f02e3ef-73cc-42e9-92a7-4366bb33e757", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1133.859924] env[61868]: DEBUG oslo_vmware.api [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316014, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060712} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.860296] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1133.861249] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-349c0179-f069-4e64-8f19-4ff3dac9c6fa {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.887216] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] 04b0079b-8f5c-4e51-8bb9-52d81273713e/04b0079b-8f5c-4e51-8bb9-52d81273713e.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1133.887598] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-67572255-8fbe-483b-a352-73d8f7cbe42e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.909909] env[61868]: DEBUG oslo_vmware.api [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1133.909909] env[61868]: value = "task-1316015" [ 1133.909909] env[61868]: _type = "Task" [ 1133.909909] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.919216] env[61868]: DEBUG oslo_vmware.api [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316015, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.947330] env[61868]: DEBUG oslo_vmware.api [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5251cc3c-f98a-314b-5cd0-a1fe1e2a9928, 'name': SearchDatastore_Task, 'duration_secs': 0.008733} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.947744] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1133.947999] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1133.948307] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1133.948576] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1133.948754] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1133.949086] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-988fcae4-aea7-4d1d-90ac-143479452cee {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.960667] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1133.960953] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1133.961783] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c6e92e8-d98a-4a6a-945b-8b1f0bcfc469 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.967626] env[61868]: DEBUG oslo_vmware.api [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 1133.967626] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52ff115f-a1b5-8edb-c7dd-86e23c2e4ef1" [ 1133.967626] env[61868]: _type = "Task" [ 1133.967626] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.978302] env[61868]: DEBUG oslo_vmware.api [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52ff115f-a1b5-8edb-c7dd-86e23c2e4ef1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.219564] env[61868]: DEBUG oslo_concurrency.lockutils [req-6c41ae39-bfe5-43a5-b18b-060f52adf237 req-527a7fbf-2834-4297-9085-439d2c007e85 service nova] Releasing lock "refresh_cache-afd7ff53-f164-4f3d-9ece-28a10245e38c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1134.422238] env[61868]: DEBUG oslo_vmware.api [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316015, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.477647] env[61868]: DEBUG oslo_vmware.api [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52ff115f-a1b5-8edb-c7dd-86e23c2e4ef1, 'name': SearchDatastore_Task, 'duration_secs': 0.008249} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.478448] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-18a96c51-6172-4b76-9d4c-ba210f10db07 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.483618] env[61868]: DEBUG oslo_vmware.api [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 1134.483618] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]525550ca-33fe-451a-86f3-2e11cbfb4864" [ 1134.483618] env[61868]: _type = "Task" [ 1134.483618] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.493054] env[61868]: DEBUG oslo_vmware.api [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]525550ca-33fe-451a-86f3-2e11cbfb4864, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.921489] env[61868]: DEBUG oslo_vmware.api [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316015, 'name': ReconfigVM_Task, 'duration_secs': 0.767963} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.921865] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Reconfigured VM instance instance-00000069 to attach disk [datastore1] 04b0079b-8f5c-4e51-8bb9-52d81273713e/04b0079b-8f5c-4e51-8bb9-52d81273713e.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1134.922602] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fe7205f1-c460-4228-9226-6cbfb43f0fe7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.929223] env[61868]: DEBUG oslo_vmware.api [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1134.929223] env[61868]: value = "task-1316016" [ 1134.929223] env[61868]: _type = "Task" [ 1134.929223] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.937715] env[61868]: DEBUG oslo_vmware.api [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316016, 'name': Rename_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.993260] env[61868]: DEBUG oslo_vmware.api [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]525550ca-33fe-451a-86f3-2e11cbfb4864, 'name': SearchDatastore_Task, 'duration_secs': 0.009209} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.993668] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1134.993816] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] afd7ff53-f164-4f3d-9ece-28a10245e38c/afd7ff53-f164-4f3d-9ece-28a10245e38c.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1134.994102] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f479e2d8-c5b6-4c21-9736-ad4d2f34327c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.000557] env[61868]: DEBUG oslo_vmware.api [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 1135.000557] env[61868]: value = "task-1316017" [ 1135.000557] env[61868]: _type = "Task" [ 1135.000557] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.008420] env[61868]: DEBUG oslo_vmware.api [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316017, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.439707] env[61868]: DEBUG oslo_vmware.api [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316016, 'name': Rename_Task, 'duration_secs': 0.16877} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.440066] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1135.440374] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7ca8529d-cc16-47fd-870d-5114860073aa {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.447913] env[61868]: DEBUG oslo_vmware.api [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1135.447913] env[61868]: value = "task-1316018" [ 1135.447913] env[61868]: _type = "Task" [ 1135.447913] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.457123] env[61868]: DEBUG oslo_vmware.api [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316018, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.510458] env[61868]: DEBUG oslo_vmware.api [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316017, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.453479} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.510756] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] afd7ff53-f164-4f3d-9ece-28a10245e38c/afd7ff53-f164-4f3d-9ece-28a10245e38c.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1135.510974] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1135.511260] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-af60b3d7-7923-4f05-9a75-a109a78f88b3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.517302] env[61868]: DEBUG oslo_vmware.api [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 1135.517302] env[61868]: value = "task-1316019" [ 1135.517302] env[61868]: _type = "Task" [ 1135.517302] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.525881] env[61868]: DEBUG oslo_vmware.api [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316019, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.958942] env[61868]: DEBUG oslo_vmware.api [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316018, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.026999] env[61868]: DEBUG oslo_vmware.api [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316019, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061535} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.027343] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1136.028129] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4343e469-6794-4631-a125-f3d47d48dde6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.050195] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] afd7ff53-f164-4f3d-9ece-28a10245e38c/afd7ff53-f164-4f3d-9ece-28a10245e38c.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1136.050480] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3c377461-d8e7-4426-b3af-6eb8fa10e9a6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.070169] env[61868]: DEBUG oslo_vmware.api [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 1136.070169] env[61868]: value = "task-1316020" [ 1136.070169] env[61868]: _type = "Task" [ 1136.070169] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.078257] env[61868]: DEBUG oslo_vmware.api [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316020, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.459151] env[61868]: DEBUG oslo_vmware.api [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316018, 'name': PowerOnVM_Task, 'duration_secs': 0.833495} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.459412] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1136.459727] env[61868]: INFO nova.compute.manager [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Took 7.65 seconds to spawn the instance on the hypervisor. [ 1136.459955] env[61868]: DEBUG nova.compute.manager [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1136.460723] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5690b7b7-b3ba-4849-b7b5-09126eca4a78 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.579168] env[61868]: DEBUG oslo_vmware.api [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316020, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.977089] env[61868]: INFO nova.compute.manager [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Took 12.34 seconds to build instance. [ 1137.080048] env[61868]: DEBUG oslo_vmware.api [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316020, 'name': ReconfigVM_Task, 'duration_secs': 0.549242} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.080356] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Reconfigured VM instance instance-0000006a to attach disk [datastore1] afd7ff53-f164-4f3d-9ece-28a10245e38c/afd7ff53-f164-4f3d-9ece-28a10245e38c.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1137.080984] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-08c22b6c-6bfd-46fe-9078-4598e817bb06 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.087131] env[61868]: DEBUG oslo_vmware.api [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 1137.087131] env[61868]: value = "task-1316021" [ 1137.087131] env[61868]: _type = "Task" [ 1137.087131] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.094303] env[61868]: DEBUG oslo_vmware.api [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316021, 'name': Rename_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.479389] env[61868]: DEBUG oslo_concurrency.lockutils [None req-14697b74-3130-45bb-8ab5-3af30876f151 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "04b0079b-8f5c-4e51-8bb9-52d81273713e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.852s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.530882] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._sync_power_states {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1137.598306] env[61868]: DEBUG oslo_vmware.api [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316021, 'name': Rename_Task, 'duration_secs': 0.27466} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.598634] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1137.598917] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-232eaba0-e733-4157-a3cb-9be3548e5cd0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.605619] env[61868]: DEBUG oslo_vmware.api [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 1137.605619] env[61868]: value = "task-1316022" [ 1137.605619] env[61868]: _type = "Task" [ 1137.605619] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.617191] env[61868]: DEBUG oslo_vmware.api [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316022, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.893055] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7a54ebb0-8dff-4b7d-91ad-7b3aacb0f928 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "04b0079b-8f5c-4e51-8bb9-52d81273713e" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.893261] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7a54ebb0-8dff-4b7d-91ad-7b3aacb0f928 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "04b0079b-8f5c-4e51-8bb9-52d81273713e" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.893490] env[61868]: DEBUG nova.compute.manager [None req-7a54ebb0-8dff-4b7d-91ad-7b3aacb0f928 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1137.894502] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d08ce99c-8c9e-4b1e-acd7-1640e471d02b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.901132] env[61868]: DEBUG nova.compute.manager [None req-7a54ebb0-8dff-4b7d-91ad-7b3aacb0f928 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61868) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1137.901692] env[61868]: DEBUG nova.objects.instance [None req-7a54ebb0-8dff-4b7d-91ad-7b3aacb0f928 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lazy-loading 'flavor' on Instance uuid 04b0079b-8f5c-4e51-8bb9-52d81273713e {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1138.035593] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Getting list of instances from cluster (obj){ [ 1138.035593] env[61868]: value = "domain-c8" [ 1138.035593] env[61868]: _type = "ClusterComputeResource" [ 1138.035593] env[61868]: } {{(pid=61868) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1138.036692] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cef904ac-f11a-431d-b787-5f88a922ffa9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.050529] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Got total of 4 instances {{(pid=61868) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1138.050760] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Triggering sync for uuid 59171bf1-f2d4-4ff5-a661-025f5dca7600 {{(pid=61868) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1138.050961] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Triggering sync for uuid f2d864f1-cd11-4b62-857d-789cf045f22c {{(pid=61868) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1138.051134] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Triggering sync for uuid 04b0079b-8f5c-4e51-8bb9-52d81273713e {{(pid=61868) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1138.051291] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Triggering sync for uuid afd7ff53-f164-4f3d-9ece-28a10245e38c {{(pid=61868) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1138.051633] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "59171bf1-f2d4-4ff5-a661-025f5dca7600" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.051859] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "59171bf1-f2d4-4ff5-a661-025f5dca7600" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.052192] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "f2d864f1-cd11-4b62-857d-789cf045f22c" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.052403] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "f2d864f1-cd11-4b62-857d-789cf045f22c" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.052603] env[61868]: INFO nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] During sync_power_state the instance has a pending task (image_uploading). Skip. [ 1138.052780] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "f2d864f1-cd11-4b62-857d-789cf045f22c" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.052979] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "04b0079b-8f5c-4e51-8bb9-52d81273713e" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.053208] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "afd7ff53-f164-4f3d-9ece-28a10245e38c" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.054083] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0823f4d7-5ddb-4042-aa06-a08fbb089fbe {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.115551] env[61868]: DEBUG oslo_vmware.api [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316022, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.411839] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a54ebb0-8dff-4b7d-91ad-7b3aacb0f928 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1138.412142] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-479887f7-4a2b-4cec-acbb-d98914d08721 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.419749] env[61868]: DEBUG oslo_vmware.api [None req-7a54ebb0-8dff-4b7d-91ad-7b3aacb0f928 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1138.419749] env[61868]: value = "task-1316023" [ 1138.419749] env[61868]: _type = "Task" [ 1138.419749] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.428790] env[61868]: DEBUG oslo_vmware.api [None req-7a54ebb0-8dff-4b7d-91ad-7b3aacb0f928 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316023, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.562847] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "59171bf1-f2d4-4ff5-a661-025f5dca7600" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.511s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.616680] env[61868]: DEBUG oslo_vmware.api [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316022, 'name': PowerOnVM_Task, 'duration_secs': 0.595326} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.616971] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1138.617200] env[61868]: INFO nova.compute.manager [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Took 7.67 seconds to spawn the instance on the hypervisor. [ 1138.617382] env[61868]: DEBUG nova.compute.manager [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1138.618152] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d5499c3-4655-4e8f-a736-ee0308623734 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.933080] env[61868]: DEBUG oslo_vmware.api [None req-7a54ebb0-8dff-4b7d-91ad-7b3aacb0f928 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316023, 'name': PowerOffVM_Task, 'duration_secs': 0.416923} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.933518] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a54ebb0-8dff-4b7d-91ad-7b3aacb0f928 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1138.933795] env[61868]: DEBUG nova.compute.manager [None req-7a54ebb0-8dff-4b7d-91ad-7b3aacb0f928 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1138.934936] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-358b8a5b-b832-4685-90d5-f083e65ea110 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.136341] env[61868]: INFO nova.compute.manager [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Took 12.82 seconds to build instance. [ 1139.449857] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7a54ebb0-8dff-4b7d-91ad-7b3aacb0f928 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "04b0079b-8f5c-4e51-8bb9-52d81273713e" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.556s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1139.450811] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "04b0079b-8f5c-4e51-8bb9-52d81273713e" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 1.398s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1139.451111] env[61868]: INFO nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] During sync_power_state the instance has a pending task (powering-off). Skip. [ 1139.451318] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "04b0079b-8f5c-4e51-8bb9-52d81273713e" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1139.561394] env[61868]: DEBUG nova.compute.manager [req-f6054e35-cf38-4f18-a638-dd691bf5901d req-0e8449b6-57a1-46bf-afe0-6870c203417f service nova] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Received event network-changed-4f02e3ef-73cc-42e9-92a7-4366bb33e757 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1139.561394] env[61868]: DEBUG nova.compute.manager [req-f6054e35-cf38-4f18-a638-dd691bf5901d req-0e8449b6-57a1-46bf-afe0-6870c203417f service nova] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Refreshing instance network info cache due to event network-changed-4f02e3ef-73cc-42e9-92a7-4366bb33e757. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1139.561472] env[61868]: DEBUG oslo_concurrency.lockutils [req-f6054e35-cf38-4f18-a638-dd691bf5901d req-0e8449b6-57a1-46bf-afe0-6870c203417f service nova] Acquiring lock "refresh_cache-afd7ff53-f164-4f3d-9ece-28a10245e38c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1139.561576] env[61868]: DEBUG oslo_concurrency.lockutils [req-f6054e35-cf38-4f18-a638-dd691bf5901d req-0e8449b6-57a1-46bf-afe0-6870c203417f service nova] Acquired lock "refresh_cache-afd7ff53-f164-4f3d-9ece-28a10245e38c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1139.561750] env[61868]: DEBUG nova.network.neutron [req-f6054e35-cf38-4f18-a638-dd691bf5901d req-0e8449b6-57a1-46bf-afe0-6870c203417f service nova] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Refreshing network info cache for port 4f02e3ef-73cc-42e9-92a7-4366bb33e757 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1139.638547] env[61868]: DEBUG oslo_concurrency.lockutils [None req-da327f2f-8835-49d5-87b2-1e0869494208 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "afd7ff53-f164-4f3d-9ece-28a10245e38c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.324s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1139.638868] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "afd7ff53-f164-4f3d-9ece-28a10245e38c" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 1.586s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1139.639079] env[61868]: INFO nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] During sync_power_state the instance has a pending task (spawning). Skip. [ 1139.639267] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "afd7ff53-f164-4f3d-9ece-28a10245e38c" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.269710] env[61868]: DEBUG nova.network.neutron [req-f6054e35-cf38-4f18-a638-dd691bf5901d req-0e8449b6-57a1-46bf-afe0-6870c203417f service nova] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Updated VIF entry in instance network info cache for port 4f02e3ef-73cc-42e9-92a7-4366bb33e757. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1140.270136] env[61868]: DEBUG nova.network.neutron [req-f6054e35-cf38-4f18-a638-dd691bf5901d req-0e8449b6-57a1-46bf-afe0-6870c203417f service nova] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Updating instance_info_cache with network_info: [{"id": "4f02e3ef-73cc-42e9-92a7-4366bb33e757", "address": "fa:16:3e:59:03:b9", "network": {"id": "3535e048-fe57-4cbc-9489-0b77484069bd", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-748518849-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.164", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c8440ce785c344d793dd513b5f008d65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f02e3ef-73", "ovs_interfaceid": "4f02e3ef-73cc-42e9-92a7-4366bb33e757", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1140.401775] env[61868]: INFO nova.compute.manager [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Rebuilding instance [ 1140.443855] env[61868]: DEBUG nova.compute.manager [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1140.445158] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9165f7a-fcbd-4a9e-b98a-49f8e6002dc9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.772863] env[61868]: DEBUG oslo_concurrency.lockutils [req-f6054e35-cf38-4f18-a638-dd691bf5901d req-0e8449b6-57a1-46bf-afe0-6870c203417f service nova] Releasing lock "refresh_cache-afd7ff53-f164-4f3d-9ece-28a10245e38c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1140.956669] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1140.957092] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fe01d4dd-b502-4b15-808b-674281b90327 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.965024] env[61868]: DEBUG oslo_vmware.api [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1140.965024] env[61868]: value = "task-1316024" [ 1140.965024] env[61868]: _type = "Task" [ 1140.965024] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.975145] env[61868]: DEBUG oslo_vmware.api [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316024, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.476135] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] VM already powered off {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1141.476402] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1141.477209] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e836f54-129d-446a-84fe-8452c06be5b1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.483777] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1141.484010] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-05851fe3-b2ae-4345-9a3a-fde39d6eafd1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.569761] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1141.570371] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1141.570371] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Deleting the datastore file [datastore1] 04b0079b-8f5c-4e51-8bb9-52d81273713e {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1141.570608] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-824e7b48-2120-444c-8712-f2a181e3b9c9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.577094] env[61868]: DEBUG oslo_vmware.api [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1141.577094] env[61868]: value = "task-1316026" [ 1141.577094] env[61868]: _type = "Task" [ 1141.577094] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.584990] env[61868]: DEBUG oslo_vmware.api [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316026, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.658821] env[61868]: DEBUG oslo_vmware.rw_handles [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52154e2c-a08a-7555-550d-8baa8c79b013/disk-0.vmdk. {{(pid=61868) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1141.659798] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36b0cc13-a988-48de-b190-a4e45023dafb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.665813] env[61868]: DEBUG oslo_vmware.rw_handles [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52154e2c-a08a-7555-550d-8baa8c79b013/disk-0.vmdk is in state: ready. {{(pid=61868) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1141.665981] env[61868]: ERROR oslo_vmware.rw_handles [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52154e2c-a08a-7555-550d-8baa8c79b013/disk-0.vmdk due to incomplete transfer. [ 1141.666205] env[61868]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-542ad21d-c86c-4d4c-968d-0ebb9683a30c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.672903] env[61868]: DEBUG oslo_vmware.rw_handles [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52154e2c-a08a-7555-550d-8baa8c79b013/disk-0.vmdk. {{(pid=61868) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1141.673118] env[61868]: DEBUG nova.virt.vmwareapi.images [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Uploaded image e70ee3c8-6ebd-4457-8ba6-ef94cd5a1e56 to the Glance image server {{(pid=61868) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1141.675227] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Destroying the VM {{(pid=61868) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1141.675474] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-ea2c7c4b-fc62-4262-9f81-93812f527f27 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.682064] env[61868]: DEBUG oslo_vmware.api [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1141.682064] env[61868]: value = "task-1316027" [ 1141.682064] env[61868]: _type = "Task" [ 1141.682064] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.689190] env[61868]: DEBUG oslo_vmware.api [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316027, 'name': Destroy_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.086302] env[61868]: DEBUG oslo_vmware.api [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316026, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.121098} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.086556] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1142.086739] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1142.086917] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1142.191391] env[61868]: DEBUG oslo_vmware.api [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316027, 'name': Destroy_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.692627] env[61868]: DEBUG oslo_vmware.api [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316027, 'name': Destroy_Task, 'duration_secs': 0.824126} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.692973] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Destroyed the VM [ 1142.693130] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Deleting Snapshot of the VM instance {{(pid=61868) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1142.693369] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-07199103-291e-40f9-aa34-ab8a11a6eb80 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.699256] env[61868]: DEBUG oslo_vmware.api [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1142.699256] env[61868]: value = "task-1316028" [ 1142.699256] env[61868]: _type = "Task" [ 1142.699256] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.706384] env[61868]: DEBUG oslo_vmware.api [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316028, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.117264] env[61868]: DEBUG nova.virt.hardware [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1143.117544] env[61868]: DEBUG nova.virt.hardware [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1143.117719] env[61868]: DEBUG nova.virt.hardware [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1143.117907] env[61868]: DEBUG nova.virt.hardware [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1143.118073] env[61868]: DEBUG nova.virt.hardware [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1143.118230] env[61868]: DEBUG nova.virt.hardware [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1143.118437] env[61868]: DEBUG nova.virt.hardware [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1143.118601] env[61868]: DEBUG nova.virt.hardware [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1143.118809] env[61868]: DEBUG nova.virt.hardware [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1143.119008] env[61868]: DEBUG nova.virt.hardware [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1143.119205] env[61868]: DEBUG nova.virt.hardware [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1143.120071] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cfe79a2-2ee3-4498-8417-17deeca0e341 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.128104] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b642add-d4f9-440d-8504-21d5810e07b1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.143498] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e7:b3:ac', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd298db54-f13d-4bf6-b6c2-755074b3047f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bf8801b5-7f14-408d-95bd-deab0e42905a', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1143.151032] env[61868]: DEBUG oslo.service.loopingcall [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1143.151285] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1143.151501] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-165b2e30-7fd1-4741-8338-2aaddb929f20 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.169686] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1143.169686] env[61868]: value = "task-1316029" [ 1143.169686] env[61868]: _type = "Task" [ 1143.169686] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.177091] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1316029, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.208185] env[61868]: DEBUG oslo_vmware.api [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316028, 'name': RemoveSnapshot_Task, 'duration_secs': 0.496219} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.208455] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Deleted Snapshot of the VM instance {{(pid=61868) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1143.208692] env[61868]: INFO nova.compute.manager [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Took 14.47 seconds to snapshot the instance on the hypervisor. [ 1143.679789] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1316029, 'name': CreateVM_Task, 'duration_secs': 0.356672} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.679963] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1143.680678] env[61868]: DEBUG oslo_concurrency.lockutils [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1143.680848] env[61868]: DEBUG oslo_concurrency.lockutils [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1143.681191] env[61868]: DEBUG oslo_concurrency.lockutils [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1143.681449] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-26fb69f6-9960-44f1-8221-9614dc834c92 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.685842] env[61868]: DEBUG oslo_vmware.api [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1143.685842] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]527cbb02-f012-d136-e5d5-a9ff3517b1a2" [ 1143.685842] env[61868]: _type = "Task" [ 1143.685842] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.693330] env[61868]: DEBUG oslo_vmware.api [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]527cbb02-f012-d136-e5d5-a9ff3517b1a2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.751574] env[61868]: DEBUG nova.compute.manager [None req-1da772c7-fd5f-4a36-85c4-e0553c107a23 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Found 1 images (rotation: 2) {{(pid=61868) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 1144.196509] env[61868]: DEBUG oslo_vmware.api [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]527cbb02-f012-d136-e5d5-a9ff3517b1a2, 'name': SearchDatastore_Task, 'duration_secs': 0.009788} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.196799] env[61868]: DEBUG oslo_concurrency.lockutils [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1144.197073] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1144.197316] env[61868]: DEBUG oslo_concurrency.lockutils [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1144.197470] env[61868]: DEBUG oslo_concurrency.lockutils [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1144.197654] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1144.197899] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4a67b942-00cc-493e-9e88-12dca1bd3b79 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.205903] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1144.206089] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1144.206773] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54f7b557-ad8a-4ea9-8e46-8d6967eb0136 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.211743] env[61868]: DEBUG oslo_vmware.api [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1144.211743] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52ae5d49-125e-194c-15c2-ab23a1f87704" [ 1144.211743] env[61868]: _type = "Task" [ 1144.211743] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.218739] env[61868]: DEBUG oslo_vmware.api [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52ae5d49-125e-194c-15c2-ab23a1f87704, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.546548] env[61868]: DEBUG nova.compute.manager [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1144.547477] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a5e16ab-9a30-4c22-a742-c3221f04c2d7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.720893] env[61868]: DEBUG oslo_vmware.api [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52ae5d49-125e-194c-15c2-ab23a1f87704, 'name': SearchDatastore_Task, 'duration_secs': 0.008941} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.721653] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8b3c0cd9-6dda-4301-986e-444396130beb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.727109] env[61868]: DEBUG oslo_vmware.api [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1144.727109] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52bf6695-65f3-789d-2063-6766ad5878d8" [ 1144.727109] env[61868]: _type = "Task" [ 1144.727109] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.734546] env[61868]: DEBUG oslo_vmware.api [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52bf6695-65f3-789d-2063-6766ad5878d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.057793] env[61868]: INFO nova.compute.manager [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] instance snapshotting [ 1145.058422] env[61868]: DEBUG nova.objects.instance [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lazy-loading 'flavor' on Instance uuid f2d864f1-cd11-4b62-857d-789cf045f22c {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1145.237587] env[61868]: DEBUG oslo_vmware.api [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52bf6695-65f3-789d-2063-6766ad5878d8, 'name': SearchDatastore_Task, 'duration_secs': 0.009225} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.237855] env[61868]: DEBUG oslo_concurrency.lockutils [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1145.238179] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 04b0079b-8f5c-4e51-8bb9-52d81273713e/04b0079b-8f5c-4e51-8bb9-52d81273713e.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1145.238442] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5c990b4e-d8e6-4419-b290-1befecd26edd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.245378] env[61868]: DEBUG oslo_vmware.api [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1145.245378] env[61868]: value = "task-1316030" [ 1145.245378] env[61868]: _type = "Task" [ 1145.245378] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.252786] env[61868]: DEBUG oslo_vmware.api [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316030, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.565190] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e31dd544-cba4-4b2a-800a-049ed9ad31f4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.585358] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea6cc003-d22a-4cc1-ab3b-abd1f7b20208 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.754507] env[61868]: DEBUG oslo_vmware.api [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316030, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.477764} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.754909] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 04b0079b-8f5c-4e51-8bb9-52d81273713e/04b0079b-8f5c-4e51-8bb9-52d81273713e.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1145.754992] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1145.755217] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-729bce1c-5147-40bc-92f4-c830d9818d04 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.761486] env[61868]: DEBUG oslo_vmware.api [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1145.761486] env[61868]: value = "task-1316031" [ 1145.761486] env[61868]: _type = "Task" [ 1145.761486] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.769182] env[61868]: DEBUG oslo_vmware.api [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316031, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.097325] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Creating Snapshot of the VM instance {{(pid=61868) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1146.098111] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-f7d71c3c-2416-440b-b028-873af474a5c1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.106918] env[61868]: DEBUG oslo_vmware.api [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1146.106918] env[61868]: value = "task-1316032" [ 1146.106918] env[61868]: _type = "Task" [ 1146.106918] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.116012] env[61868]: DEBUG oslo_vmware.api [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316032, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.271586] env[61868]: DEBUG oslo_vmware.api [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316031, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.055414} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.271864] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1146.272647] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab8231c5-77f9-4f2e-9702-fa9ecb266a0e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.294125] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] 04b0079b-8f5c-4e51-8bb9-52d81273713e/04b0079b-8f5c-4e51-8bb9-52d81273713e.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1146.294385] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c2d00cc7-b64f-430e-8abc-936da0cf6e9e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.312756] env[61868]: DEBUG oslo_vmware.api [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1146.312756] env[61868]: value = "task-1316033" [ 1146.312756] env[61868]: _type = "Task" [ 1146.312756] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.320086] env[61868]: DEBUG oslo_vmware.api [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316033, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.618177] env[61868]: DEBUG oslo_vmware.api [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316032, 'name': CreateSnapshot_Task, 'duration_secs': 0.433513} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.618457] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Created Snapshot of the VM instance {{(pid=61868) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1146.619173] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aba90809-4b57-4da6-b642-57b693474121 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.821911] env[61868]: DEBUG oslo_vmware.api [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316033, 'name': ReconfigVM_Task, 'duration_secs': 0.269201} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.822275] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Reconfigured VM instance instance-00000069 to attach disk [datastore1] 04b0079b-8f5c-4e51-8bb9-52d81273713e/04b0079b-8f5c-4e51-8bb9-52d81273713e.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1146.822743] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5515ccb3-58a6-409b-9751-639ba7e878d8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.828719] env[61868]: DEBUG oslo_vmware.api [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1146.828719] env[61868]: value = "task-1316034" [ 1146.828719] env[61868]: _type = "Task" [ 1146.828719] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.836855] env[61868]: DEBUG oslo_vmware.api [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316034, 'name': Rename_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.135672] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Creating linked-clone VM from snapshot {{(pid=61868) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1147.136021] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-1929a958-5c61-44fe-a0a0-9d50f30870e7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.144195] env[61868]: DEBUG oslo_vmware.api [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1147.144195] env[61868]: value = "task-1316035" [ 1147.144195] env[61868]: _type = "Task" [ 1147.144195] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.152881] env[61868]: DEBUG oslo_vmware.api [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316035, 'name': CloneVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.342227] env[61868]: DEBUG oslo_vmware.api [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316034, 'name': Rename_Task, 'duration_secs': 0.144685} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.342647] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1147.343109] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c28ebcf6-0785-4f93-8b95-c575a7d0c8d5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.349875] env[61868]: DEBUG oslo_vmware.api [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1147.349875] env[61868]: value = "task-1316036" [ 1147.349875] env[61868]: _type = "Task" [ 1147.349875] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.357579] env[61868]: DEBUG oslo_vmware.api [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316036, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.655103] env[61868]: DEBUG oslo_vmware.api [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316035, 'name': CloneVM_Task} progress is 94%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.859569] env[61868]: DEBUG oslo_vmware.api [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316036, 'name': PowerOnVM_Task, 'duration_secs': 0.421954} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.860049] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1147.860049] env[61868]: DEBUG nova.compute.manager [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1147.860787] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cedf8f3-b4c6-4a0a-864c-78bd7e94032e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.155992] env[61868]: DEBUG oslo_vmware.api [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316035, 'name': CloneVM_Task} progress is 95%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.371701] env[61868]: INFO nova.compute.manager [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] bringing vm to original state: 'stopped' [ 1148.659068] env[61868]: DEBUG oslo_vmware.api [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316035, 'name': CloneVM_Task, 'duration_secs': 1.286298} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.659437] env[61868]: INFO nova.virt.vmwareapi.vmops [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Created linked-clone VM from snapshot [ 1148.660461] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec2f84bb-27ef-4532-9a2d-a535d95803a8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.670265] env[61868]: DEBUG nova.virt.vmwareapi.images [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Uploading image 3848f10c-0a5e-4afe-81d9-129af750dbe5 {{(pid=61868) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1148.755428] env[61868]: DEBUG oslo_vmware.rw_handles [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1148.755428] env[61868]: value = "vm-281634" [ 1148.755428] env[61868]: _type = "VirtualMachine" [ 1148.755428] env[61868]: }. {{(pid=61868) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1148.755741] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-e0db8b5b-9743-4852-a741-fe94c2e8a89c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.762459] env[61868]: DEBUG oslo_vmware.rw_handles [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lease: (returnval){ [ 1148.762459] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52ce8fd8-1e99-8f38-0cc8-720ee4c648e3" [ 1148.762459] env[61868]: _type = "HttpNfcLease" [ 1148.762459] env[61868]: } obtained for exporting VM: (result){ [ 1148.762459] env[61868]: value = "vm-281634" [ 1148.762459] env[61868]: _type = "VirtualMachine" [ 1148.762459] env[61868]: }. {{(pid=61868) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1148.762688] env[61868]: DEBUG oslo_vmware.api [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the lease: (returnval){ [ 1148.762688] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52ce8fd8-1e99-8f38-0cc8-720ee4c648e3" [ 1148.762688] env[61868]: _type = "HttpNfcLease" [ 1148.762688] env[61868]: } to be ready. {{(pid=61868) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1148.768820] env[61868]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1148.768820] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52ce8fd8-1e99-8f38-0cc8-720ee4c648e3" [ 1148.768820] env[61868]: _type = "HttpNfcLease" [ 1148.768820] env[61868]: } is initializing. {{(pid=61868) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1149.270516] env[61868]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1149.270516] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52ce8fd8-1e99-8f38-0cc8-720ee4c648e3" [ 1149.270516] env[61868]: _type = "HttpNfcLease" [ 1149.270516] env[61868]: } is ready. {{(pid=61868) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1149.270951] env[61868]: DEBUG oslo_vmware.rw_handles [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1149.270951] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52ce8fd8-1e99-8f38-0cc8-720ee4c648e3" [ 1149.270951] env[61868]: _type = "HttpNfcLease" [ 1149.270951] env[61868]: }. {{(pid=61868) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1149.271527] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64d6f2f7-3c09-4eaf-9d24-07521835bf91 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.278404] env[61868]: DEBUG oslo_vmware.rw_handles [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528cd94e-31e6-b4f9-edb0-c1efe0fa4b6c/disk-0.vmdk from lease info. {{(pid=61868) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1149.278577] env[61868]: DEBUG oslo_vmware.rw_handles [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528cd94e-31e6-b4f9-edb0-c1efe0fa4b6c/disk-0.vmdk for reading. {{(pid=61868) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1149.366250] env[61868]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d533aab4-9e41-4476-9662-30c8c810e570 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.380346] env[61868]: DEBUG oslo_concurrency.lockutils [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "04b0079b-8f5c-4e51-8bb9-52d81273713e" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1149.380660] env[61868]: DEBUG oslo_concurrency.lockutils [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "04b0079b-8f5c-4e51-8bb9-52d81273713e" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1149.380861] env[61868]: DEBUG nova.compute.manager [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1149.381887] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-753f6e93-56d2-4872-b2cb-bea6872d449b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.388806] env[61868]: DEBUG nova.compute.manager [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61868) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1149.391091] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1149.391317] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4987c057-a2fe-44b3-8354-c3ef1a87af79 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.405106] env[61868]: DEBUG oslo_vmware.api [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1149.405106] env[61868]: value = "task-1316038" [ 1149.405106] env[61868]: _type = "Task" [ 1149.405106] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.412985] env[61868]: DEBUG oslo_vmware.api [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316038, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.915722] env[61868]: DEBUG oslo_vmware.api [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316038, 'name': PowerOffVM_Task, 'duration_secs': 0.156808} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.916280] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1149.917175] env[61868]: DEBUG nova.compute.manager [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1149.917291] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92489729-7113-43d7-919c-92cef14ec7e5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.430300] env[61868]: DEBUG oslo_concurrency.lockutils [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "04b0079b-8f5c-4e51-8bb9-52d81273713e" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.050s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1150.939037] env[61868]: DEBUG oslo_concurrency.lockutils [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1150.939037] env[61868]: DEBUG oslo_concurrency.lockutils [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1150.939037] env[61868]: DEBUG nova.objects.instance [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61868) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1151.881895] env[61868]: DEBUG oslo_concurrency.lockutils [None req-efc294a3-00fe-4f0f-a23e-bc473cc95aa9 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "04b0079b-8f5c-4e51-8bb9-52d81273713e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1151.882287] env[61868]: DEBUG oslo_concurrency.lockutils [None req-efc294a3-00fe-4f0f-a23e-bc473cc95aa9 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "04b0079b-8f5c-4e51-8bb9-52d81273713e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1151.882443] env[61868]: DEBUG oslo_concurrency.lockutils [None req-efc294a3-00fe-4f0f-a23e-bc473cc95aa9 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "04b0079b-8f5c-4e51-8bb9-52d81273713e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1151.882638] env[61868]: DEBUG oslo_concurrency.lockutils [None req-efc294a3-00fe-4f0f-a23e-bc473cc95aa9 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "04b0079b-8f5c-4e51-8bb9-52d81273713e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1151.882832] env[61868]: DEBUG oslo_concurrency.lockutils [None req-efc294a3-00fe-4f0f-a23e-bc473cc95aa9 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "04b0079b-8f5c-4e51-8bb9-52d81273713e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1151.885194] env[61868]: INFO nova.compute.manager [None req-efc294a3-00fe-4f0f-a23e-bc473cc95aa9 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Terminating instance [ 1151.887007] env[61868]: DEBUG nova.compute.manager [None req-efc294a3-00fe-4f0f-a23e-bc473cc95aa9 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1151.887227] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-efc294a3-00fe-4f0f-a23e-bc473cc95aa9 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1151.888061] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a0ae87b-332d-4887-869d-358ea576cd3f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.895846] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-efc294a3-00fe-4f0f-a23e-bc473cc95aa9 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1151.896108] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-00839129-b5ce-4741-b6dc-59b80e313eff {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.947578] env[61868]: DEBUG oslo_concurrency.lockutils [None req-66a998c0-f8a8-4000-851e-bd15b018137f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1151.963287] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-efc294a3-00fe-4f0f-a23e-bc473cc95aa9 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1151.963538] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-efc294a3-00fe-4f0f-a23e-bc473cc95aa9 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1151.963729] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-efc294a3-00fe-4f0f-a23e-bc473cc95aa9 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Deleting the datastore file [datastore1] 04b0079b-8f5c-4e51-8bb9-52d81273713e {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1151.963998] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8661baaf-b6d7-4c06-b904-48cc3e23da3d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.970767] env[61868]: DEBUG oslo_vmware.api [None req-efc294a3-00fe-4f0f-a23e-bc473cc95aa9 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1151.970767] env[61868]: value = "task-1316040" [ 1151.970767] env[61868]: _type = "Task" [ 1151.970767] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.979127] env[61868]: DEBUG oslo_vmware.api [None req-efc294a3-00fe-4f0f-a23e-bc473cc95aa9 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316040, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.074979] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1152.075216] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61868) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1152.480476] env[61868]: DEBUG oslo_vmware.api [None req-efc294a3-00fe-4f0f-a23e-bc473cc95aa9 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316040, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152132} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.480685] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-efc294a3-00fe-4f0f-a23e-bc473cc95aa9 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1152.480895] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-efc294a3-00fe-4f0f-a23e-bc473cc95aa9 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1152.481107] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-efc294a3-00fe-4f0f-a23e-bc473cc95aa9 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1152.481307] env[61868]: INFO nova.compute.manager [None req-efc294a3-00fe-4f0f-a23e-bc473cc95aa9 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Took 0.59 seconds to destroy the instance on the hypervisor. [ 1152.481555] env[61868]: DEBUG oslo.service.loopingcall [None req-efc294a3-00fe-4f0f-a23e-bc473cc95aa9 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1152.481753] env[61868]: DEBUG nova.compute.manager [-] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1152.481868] env[61868]: DEBUG nova.network.neutron [-] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1152.780964] env[61868]: DEBUG nova.compute.manager [req-57b70abc-40df-448b-b906-dfa1843c7bd7 req-b339d2b3-733f-480f-8c8d-fff8f664ea7b service nova] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Received event network-vif-deleted-bf8801b5-7f14-408d-95bd-deab0e42905a {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1152.780964] env[61868]: INFO nova.compute.manager [req-57b70abc-40df-448b-b906-dfa1843c7bd7 req-b339d2b3-733f-480f-8c8d-fff8f664ea7b service nova] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Neutron deleted interface bf8801b5-7f14-408d-95bd-deab0e42905a; detaching it from the instance and deleting it from the info cache [ 1152.780964] env[61868]: DEBUG nova.network.neutron [req-57b70abc-40df-448b-b906-dfa1843c7bd7 req-b339d2b3-733f-480f-8c8d-fff8f664ea7b service nova] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1153.075782] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1153.223451] env[61868]: DEBUG nova.network.neutron [-] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1153.283953] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3bf582db-61c6-4747-9605-b30607452db5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.293787] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c786aa5-290b-4b9c-b8ec-4b2c0d82a12e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.320785] env[61868]: DEBUG nova.compute.manager [req-57b70abc-40df-448b-b906-dfa1843c7bd7 req-b339d2b3-733f-480f-8c8d-fff8f664ea7b service nova] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Detach interface failed, port_id=bf8801b5-7f14-408d-95bd-deab0e42905a, reason: Instance 04b0079b-8f5c-4e51-8bb9-52d81273713e could not be found. {{(pid=61868) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1153.726679] env[61868]: INFO nova.compute.manager [-] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Took 1.24 seconds to deallocate network for instance. [ 1154.233354] env[61868]: DEBUG oslo_concurrency.lockutils [None req-efc294a3-00fe-4f0f-a23e-bc473cc95aa9 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1154.233742] env[61868]: DEBUG oslo_concurrency.lockutils [None req-efc294a3-00fe-4f0f-a23e-bc473cc95aa9 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1154.233831] env[61868]: DEBUG nova.objects.instance [None req-efc294a3-00fe-4f0f-a23e-bc473cc95aa9 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lazy-loading 'resources' on Instance uuid 04b0079b-8f5c-4e51-8bb9-52d81273713e {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1154.803460] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a868cd18-20e5-468d-858b-030746812742 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.811215] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2812ee63-d5e1-456a-88a0-9fde658e20e6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.842009] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8e93bfd-7b3a-4db3-b209-49411860d54a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.850149] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f49aed8-5af8-43b6-a4e9-800d8a2e8c87 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.864383] env[61868]: DEBUG nova.compute.provider_tree [None req-efc294a3-00fe-4f0f-a23e-bc473cc95aa9 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1155.075128] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager.update_available_resource {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1155.367906] env[61868]: DEBUG nova.scheduler.client.report [None req-efc294a3-00fe-4f0f-a23e-bc473cc95aa9 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1155.578132] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.873205] env[61868]: DEBUG oslo_concurrency.lockutils [None req-efc294a3-00fe-4f0f-a23e-bc473cc95aa9 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.639s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.875801] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.298s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.876875] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.876875] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61868) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1155.877252] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69eb09ed-3858-4677-8fea-796768074931 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.886256] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef78524b-5084-4fee-846f-b252f370de14 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.900696] env[61868]: INFO nova.scheduler.client.report [None req-efc294a3-00fe-4f0f-a23e-bc473cc95aa9 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Deleted allocations for instance 04b0079b-8f5c-4e51-8bb9-52d81273713e [ 1155.902193] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad290e54-993a-4f3e-b9f2-8216e36b6569 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.910904] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49b62c36-e55a-4104-b52c-e59649923b04 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.940823] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181066MB free_disk=134GB free_vcpus=48 pci_devices=None {{(pid=61868) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1155.941032] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.941208] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1156.410946] env[61868]: DEBUG oslo_concurrency.lockutils [None req-efc294a3-00fe-4f0f-a23e-bc473cc95aa9 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "04b0079b-8f5c-4e51-8bb9-52d81273713e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.529s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1156.750319] env[61868]: DEBUG oslo_vmware.rw_handles [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528cd94e-31e6-b4f9-edb0-c1efe0fa4b6c/disk-0.vmdk. {{(pid=61868) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1156.751311] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1e693b4-667e-449d-a888-88225843f4bd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.757544] env[61868]: DEBUG oslo_vmware.rw_handles [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528cd94e-31e6-b4f9-edb0-c1efe0fa4b6c/disk-0.vmdk is in state: ready. {{(pid=61868) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1156.757713] env[61868]: ERROR oslo_vmware.rw_handles [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528cd94e-31e6-b4f9-edb0-c1efe0fa4b6c/disk-0.vmdk due to incomplete transfer. [ 1156.757931] env[61868]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-00947f21-4e98-4142-82cd-02fdd53bc900 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.764973] env[61868]: DEBUG oslo_vmware.rw_handles [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528cd94e-31e6-b4f9-edb0-c1efe0fa4b6c/disk-0.vmdk. {{(pid=61868) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1156.765186] env[61868]: DEBUG nova.virt.vmwareapi.images [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Uploaded image 3848f10c-0a5e-4afe-81d9-129af750dbe5 to the Glance image server {{(pid=61868) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1156.767504] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Destroying the VM {{(pid=61868) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1156.767731] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-2f9c034f-5755-4541-8d8c-14abe4c4e63d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.773012] env[61868]: DEBUG oslo_vmware.api [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1156.773012] env[61868]: value = "task-1316041" [ 1156.773012] env[61868]: _type = "Task" [ 1156.773012] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.780167] env[61868]: DEBUG oslo_vmware.api [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316041, 'name': Destroy_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.964969] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 59171bf1-f2d4-4ff5-a661-025f5dca7600 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1156.964969] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance f2d864f1-cd11-4b62-857d-789cf045f22c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1156.964969] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance afd7ff53-f164-4f3d-9ece-28a10245e38c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1156.964969] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=61868) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1156.964969] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=61868) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1157.010243] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a69c33e4-3595-4da2-8f83-6df6ad9bbc8d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.017882] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7ffda3f-eb2c-4ba0-8967-aa28b7a84f9e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.047027] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83eef011-7d86-49ad-9e6c-ec9a8605bb36 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.054482] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f93a8cfd-8c14-4a31-a170-4af006f67247 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.069180] env[61868]: DEBUG nova.compute.provider_tree [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1157.284407] env[61868]: DEBUG oslo_vmware.api [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316041, 'name': Destroy_Task, 'duration_secs': 0.329357} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.284778] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Destroyed the VM [ 1157.285048] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Deleting Snapshot of the VM instance {{(pid=61868) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1157.285352] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-6288ba77-b3c2-4d64-a6dd-885566b98097 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.292135] env[61868]: DEBUG oslo_vmware.api [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1157.292135] env[61868]: value = "task-1316042" [ 1157.292135] env[61868]: _type = "Task" [ 1157.292135] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.299596] env[61868]: DEBUG oslo_vmware.api [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316042, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.572802] env[61868]: DEBUG nova.scheduler.client.report [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1157.804135] env[61868]: DEBUG oslo_vmware.api [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316042, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.078799] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61868) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1158.079069] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.138s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1158.303310] env[61868]: DEBUG oslo_vmware.api [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316042, 'name': RemoveSnapshot_Task, 'duration_secs': 0.522308} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.303607] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Deleted Snapshot of the VM instance {{(pid=61868) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1158.303842] env[61868]: INFO nova.compute.manager [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Took 12.74 seconds to snapshot the instance on the hypervisor. [ 1158.779270] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1158.779513] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1158.859237] env[61868]: DEBUG nova.compute.manager [None req-3b51fd01-d486-4606-894c-eb0efad03c1d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Found 2 images (rotation: 2) {{(pid=61868) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 1159.079052] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1159.079278] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1159.079425] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Starting heal instance info cache {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1159.079545] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Rebuilding the list of instances to heal {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1159.281554] env[61868]: DEBUG nova.compute.manager [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1159.429512] env[61868]: DEBUG nova.compute.manager [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1159.430519] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9186d4e8-1871-4410-87af-834288e00d4c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.638858] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "refresh_cache-59171bf1-f2d4-4ff5-a661-025f5dca7600" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1159.639019] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquired lock "refresh_cache-59171bf1-f2d4-4ff5-a661-025f5dca7600" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1159.639183] env[61868]: DEBUG nova.network.neutron [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Forcefully refreshing network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1159.639339] env[61868]: DEBUG nova.objects.instance [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lazy-loading 'info_cache' on Instance uuid 59171bf1-f2d4-4ff5-a661-025f5dca7600 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1159.803660] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1159.803909] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1159.805440] env[61868]: INFO nova.compute.claims [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1159.942391] env[61868]: INFO nova.compute.manager [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] instance snapshotting [ 1159.943080] env[61868]: DEBUG nova.objects.instance [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lazy-loading 'flavor' on Instance uuid f2d864f1-cd11-4b62-857d-789cf045f22c {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1160.448328] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10881b49-d455-4f7a-b1e7-4b38a35cf25e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.467373] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01b42994-2e55-43c0-aabf-d59e783e77af {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.869607] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bd6509a-6eed-47bc-90f1-a1157e9efa70 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.876406] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db926604-0c7c-4476-a973-1f1c988963f3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.905666] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f62ee49-50e3-4c82-bd90-6b405e63918f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.912168] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b67aaae-2db4-49a7-833f-62bcca62e8f8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.924407] env[61868]: DEBUG nova.compute.provider_tree [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1160.976882] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Creating Snapshot of the VM instance {{(pid=61868) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1160.977177] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-465ac028-e21e-4121-9411-179cfe915e9f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.984052] env[61868]: DEBUG oslo_vmware.api [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1160.984052] env[61868]: value = "task-1316043" [ 1160.984052] env[61868]: _type = "Task" [ 1160.984052] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.990984] env[61868]: DEBUG oslo_vmware.api [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316043, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.372182] env[61868]: DEBUG nova.network.neutron [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Updating instance_info_cache with network_info: [{"id": "8f9ad20f-f367-4826-8aae-00a1192d0bc8", "address": "fa:16:3e:63:2a:68", "network": {"id": "16b3638c-1129-49b8-9f22-636ff7db7386", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1544179589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.190", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bdab2203834c4ee5bab11aca70e48cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d298db54-f13d-4bf6-b6c2-755074b3047f", "external-id": "nsx-vlan-transportzone-631", "segmentation_id": 631, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f9ad20f-f3", "ovs_interfaceid": "8f9ad20f-f367-4826-8aae-00a1192d0bc8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1161.428064] env[61868]: DEBUG nova.scheduler.client.report [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1161.493965] env[61868]: DEBUG oslo_vmware.api [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316043, 'name': CreateSnapshot_Task, 'duration_secs': 0.421458} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.494210] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Created Snapshot of the VM instance {{(pid=61868) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1161.494960] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfb0e339-6e33-4536-bfea-0f96890055c4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.873555] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Releasing lock "refresh_cache-59171bf1-f2d4-4ff5-a661-025f5dca7600" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1161.873916] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Updated the network info_cache for instance {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1161.874026] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1161.874151] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1161.874305] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1161.874447] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1161.932488] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.128s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1161.933031] env[61868]: DEBUG nova.compute.manager [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1162.011293] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Creating linked-clone VM from snapshot {{(pid=61868) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1162.011606] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-ff44e423-3d15-4e56-a13d-16a8d961ca39 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.019550] env[61868]: DEBUG oslo_vmware.api [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1162.019550] env[61868]: value = "task-1316044" [ 1162.019550] env[61868]: _type = "Task" [ 1162.019550] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.028208] env[61868]: DEBUG oslo_vmware.api [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316044, 'name': CloneVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.437976] env[61868]: DEBUG nova.compute.utils [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1162.439549] env[61868]: DEBUG nova.compute.manager [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1162.439864] env[61868]: DEBUG nova.network.neutron [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1162.492072] env[61868]: DEBUG nova.policy [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '11a64f7a464b45c192edd10486c61138', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bdab2203834c4ee5bab11aca70e48cc1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 1162.530534] env[61868]: DEBUG oslo_vmware.api [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316044, 'name': CloneVM_Task} progress is 94%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.774964] env[61868]: DEBUG nova.network.neutron [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Successfully created port: 2bac9dd8-51e8-4aad-bf36-90e045aa9106 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1162.944204] env[61868]: DEBUG nova.compute.manager [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1163.029315] env[61868]: DEBUG oslo_vmware.api [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316044, 'name': CloneVM_Task, 'duration_secs': 0.92622} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.029532] env[61868]: INFO nova.virt.vmwareapi.vmops [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Created linked-clone VM from snapshot [ 1163.030268] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-287f6a29-1ac6-445c-9af5-9afc9d9d7c89 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.037711] env[61868]: DEBUG nova.virt.vmwareapi.images [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Uploading image 456023ac-572a-4238-9427-5a28e4383daf {{(pid=61868) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1163.059136] env[61868]: DEBUG oslo_vmware.rw_handles [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1163.059136] env[61868]: value = "vm-281636" [ 1163.059136] env[61868]: _type = "VirtualMachine" [ 1163.059136] env[61868]: }. {{(pid=61868) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1163.059390] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-b7365c50-fb24-4962-a1d9-114a77825a7a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.066008] env[61868]: DEBUG oslo_vmware.rw_handles [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lease: (returnval){ [ 1163.066008] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52d7a2ae-5fcb-c60b-cd34-0ca815e49f89" [ 1163.066008] env[61868]: _type = "HttpNfcLease" [ 1163.066008] env[61868]: } obtained for exporting VM: (result){ [ 1163.066008] env[61868]: value = "vm-281636" [ 1163.066008] env[61868]: _type = "VirtualMachine" [ 1163.066008] env[61868]: }. {{(pid=61868) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1163.066257] env[61868]: DEBUG oslo_vmware.api [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the lease: (returnval){ [ 1163.066257] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52d7a2ae-5fcb-c60b-cd34-0ca815e49f89" [ 1163.066257] env[61868]: _type = "HttpNfcLease" [ 1163.066257] env[61868]: } to be ready. {{(pid=61868) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1163.072647] env[61868]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1163.072647] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52d7a2ae-5fcb-c60b-cd34-0ca815e49f89" [ 1163.072647] env[61868]: _type = "HttpNfcLease" [ 1163.072647] env[61868]: } is initializing. {{(pid=61868) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1163.574466] env[61868]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1163.574466] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52d7a2ae-5fcb-c60b-cd34-0ca815e49f89" [ 1163.574466] env[61868]: _type = "HttpNfcLease" [ 1163.574466] env[61868]: } is ready. {{(pid=61868) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1163.574801] env[61868]: DEBUG oslo_vmware.rw_handles [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1163.574801] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52d7a2ae-5fcb-c60b-cd34-0ca815e49f89" [ 1163.574801] env[61868]: _type = "HttpNfcLease" [ 1163.574801] env[61868]: }. {{(pid=61868) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1163.575582] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cab428fb-9f90-447e-b5c9-5cafd70f92b4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.582799] env[61868]: DEBUG oslo_vmware.rw_handles [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5298676d-a3c0-f2b8-d97e-6a1089bfbc9b/disk-0.vmdk from lease info. {{(pid=61868) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1163.582985] env[61868]: DEBUG oslo_vmware.rw_handles [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5298676d-a3c0-f2b8-d97e-6a1089bfbc9b/disk-0.vmdk for reading. {{(pid=61868) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1163.667442] env[61868]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-589b6882-cadd-4a15-aad6-7142d53bc932 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.954576] env[61868]: DEBUG nova.compute.manager [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1163.977436] env[61868]: DEBUG nova.virt.hardware [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1163.977722] env[61868]: DEBUG nova.virt.hardware [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1163.977883] env[61868]: DEBUG nova.virt.hardware [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1163.978104] env[61868]: DEBUG nova.virt.hardware [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1163.978259] env[61868]: DEBUG nova.virt.hardware [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1163.978410] env[61868]: DEBUG nova.virt.hardware [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1163.978637] env[61868]: DEBUG nova.virt.hardware [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1163.978808] env[61868]: DEBUG nova.virt.hardware [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1163.978983] env[61868]: DEBUG nova.virt.hardware [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1163.979175] env[61868]: DEBUG nova.virt.hardware [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1163.979352] env[61868]: DEBUG nova.virt.hardware [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1163.980294] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b23cab5-d654-4d3a-bee4-8682d7241185 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.988451] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63edcd67-36fa-497f-9927-38f81f5b6cd5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.177217] env[61868]: DEBUG nova.compute.manager [req-77574002-fa1c-482a-bb17-2fb171ccdfba req-ec61822f-d6e1-4f4b-a36f-44d91acf0ba3 service nova] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Received event network-vif-plugged-2bac9dd8-51e8-4aad-bf36-90e045aa9106 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1164.177478] env[61868]: DEBUG oslo_concurrency.lockutils [req-77574002-fa1c-482a-bb17-2fb171ccdfba req-ec61822f-d6e1-4f4b-a36f-44d91acf0ba3 service nova] Acquiring lock "7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1164.177949] env[61868]: DEBUG oslo_concurrency.lockutils [req-77574002-fa1c-482a-bb17-2fb171ccdfba req-ec61822f-d6e1-4f4b-a36f-44d91acf0ba3 service nova] Lock "7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1164.178287] env[61868]: DEBUG oslo_concurrency.lockutils [req-77574002-fa1c-482a-bb17-2fb171ccdfba req-ec61822f-d6e1-4f4b-a36f-44d91acf0ba3 service nova] Lock "7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1164.178658] env[61868]: DEBUG nova.compute.manager [req-77574002-fa1c-482a-bb17-2fb171ccdfba req-ec61822f-d6e1-4f4b-a36f-44d91acf0ba3 service nova] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] No waiting events found dispatching network-vif-plugged-2bac9dd8-51e8-4aad-bf36-90e045aa9106 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1164.178989] env[61868]: WARNING nova.compute.manager [req-77574002-fa1c-482a-bb17-2fb171ccdfba req-ec61822f-d6e1-4f4b-a36f-44d91acf0ba3 service nova] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Received unexpected event network-vif-plugged-2bac9dd8-51e8-4aad-bf36-90e045aa9106 for instance with vm_state building and task_state spawning. [ 1164.269060] env[61868]: DEBUG nova.network.neutron [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Successfully updated port: 2bac9dd8-51e8-4aad-bf36-90e045aa9106 {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1164.772300] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "refresh_cache-7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1164.772495] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquired lock "refresh_cache-7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1164.772589] env[61868]: DEBUG nova.network.neutron [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1165.305586] env[61868]: DEBUG nova.network.neutron [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1165.440549] env[61868]: DEBUG nova.network.neutron [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Updating instance_info_cache with network_info: [{"id": "2bac9dd8-51e8-4aad-bf36-90e045aa9106", "address": "fa:16:3e:d7:73:72", "network": {"id": "16b3638c-1129-49b8-9f22-636ff7db7386", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1544179589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bdab2203834c4ee5bab11aca70e48cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d298db54-f13d-4bf6-b6c2-755074b3047f", "external-id": "nsx-vlan-transportzone-631", "segmentation_id": 631, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2bac9dd8-51", "ovs_interfaceid": "2bac9dd8-51e8-4aad-bf36-90e045aa9106", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1165.943535] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Releasing lock "refresh_cache-7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1165.943912] env[61868]: DEBUG nova.compute.manager [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Instance network_info: |[{"id": "2bac9dd8-51e8-4aad-bf36-90e045aa9106", "address": "fa:16:3e:d7:73:72", "network": {"id": "16b3638c-1129-49b8-9f22-636ff7db7386", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1544179589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bdab2203834c4ee5bab11aca70e48cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d298db54-f13d-4bf6-b6c2-755074b3047f", "external-id": "nsx-vlan-transportzone-631", "segmentation_id": 631, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2bac9dd8-51", "ovs_interfaceid": "2bac9dd8-51e8-4aad-bf36-90e045aa9106", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1165.944416] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d7:73:72', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd298db54-f13d-4bf6-b6c2-755074b3047f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2bac9dd8-51e8-4aad-bf36-90e045aa9106', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1165.951983] env[61868]: DEBUG oslo.service.loopingcall [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1165.954165] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1165.954518] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fa3e0380-9585-4157-87fa-f0aee518e780 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.973282] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1165.973282] env[61868]: value = "task-1316046" [ 1165.973282] env[61868]: _type = "Task" [ 1165.973282] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.981179] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1316046, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.206224] env[61868]: DEBUG nova.compute.manager [req-31155f82-f33e-48bb-a2c7-7ad1491ef795 req-e28c8b35-b0d9-41a9-8274-bb3cf29a6462 service nova] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Received event network-changed-2bac9dd8-51e8-4aad-bf36-90e045aa9106 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1166.206497] env[61868]: DEBUG nova.compute.manager [req-31155f82-f33e-48bb-a2c7-7ad1491ef795 req-e28c8b35-b0d9-41a9-8274-bb3cf29a6462 service nova] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Refreshing instance network info cache due to event network-changed-2bac9dd8-51e8-4aad-bf36-90e045aa9106. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1166.206724] env[61868]: DEBUG oslo_concurrency.lockutils [req-31155f82-f33e-48bb-a2c7-7ad1491ef795 req-e28c8b35-b0d9-41a9-8274-bb3cf29a6462 service nova] Acquiring lock "refresh_cache-7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1166.206875] env[61868]: DEBUG oslo_concurrency.lockutils [req-31155f82-f33e-48bb-a2c7-7ad1491ef795 req-e28c8b35-b0d9-41a9-8274-bb3cf29a6462 service nova] Acquired lock "refresh_cache-7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1166.207054] env[61868]: DEBUG nova.network.neutron [req-31155f82-f33e-48bb-a2c7-7ad1491ef795 req-e28c8b35-b0d9-41a9-8274-bb3cf29a6462 service nova] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Refreshing network info cache for port 2bac9dd8-51e8-4aad-bf36-90e045aa9106 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1166.483578] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1316046, 'name': CreateVM_Task, 'duration_secs': 0.3229} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.483917] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1166.484450] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1166.484639] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1166.485015] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1166.485289] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f14e7e23-f5c5-4bec-a83f-ec0edf084214 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.489616] env[61868]: DEBUG oslo_vmware.api [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1166.489616] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5293a712-12df-ce26-8d80-d0548b54f598" [ 1166.489616] env[61868]: _type = "Task" [ 1166.489616] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.497288] env[61868]: DEBUG oslo_vmware.api [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5293a712-12df-ce26-8d80-d0548b54f598, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.909345] env[61868]: DEBUG nova.network.neutron [req-31155f82-f33e-48bb-a2c7-7ad1491ef795 req-e28c8b35-b0d9-41a9-8274-bb3cf29a6462 service nova] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Updated VIF entry in instance network info cache for port 2bac9dd8-51e8-4aad-bf36-90e045aa9106. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1166.909916] env[61868]: DEBUG nova.network.neutron [req-31155f82-f33e-48bb-a2c7-7ad1491ef795 req-e28c8b35-b0d9-41a9-8274-bb3cf29a6462 service nova] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Updating instance_info_cache with network_info: [{"id": "2bac9dd8-51e8-4aad-bf36-90e045aa9106", "address": "fa:16:3e:d7:73:72", "network": {"id": "16b3638c-1129-49b8-9f22-636ff7db7386", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1544179589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bdab2203834c4ee5bab11aca70e48cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d298db54-f13d-4bf6-b6c2-755074b3047f", "external-id": "nsx-vlan-transportzone-631", "segmentation_id": 631, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2bac9dd8-51", "ovs_interfaceid": "2bac9dd8-51e8-4aad-bf36-90e045aa9106", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1167.003848] env[61868]: DEBUG oslo_vmware.api [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5293a712-12df-ce26-8d80-d0548b54f598, 'name': SearchDatastore_Task, 'duration_secs': 0.012277} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.004214] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1167.004548] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1167.004887] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1167.005119] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1167.005386] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1167.005740] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-08ef4b09-2b38-413a-bdc9-55067648587b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.016324] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1167.016586] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1167.017671] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e372f8c-fa3f-4bde-bace-60cb27dd6e36 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.024312] env[61868]: DEBUG oslo_vmware.api [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1167.024312] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5277061c-dfa6-af4e-a626-d75a55c665a3" [ 1167.024312] env[61868]: _type = "Task" [ 1167.024312] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.034233] env[61868]: DEBUG oslo_vmware.api [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5277061c-dfa6-af4e-a626-d75a55c665a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.413640] env[61868]: DEBUG oslo_concurrency.lockutils [req-31155f82-f33e-48bb-a2c7-7ad1491ef795 req-e28c8b35-b0d9-41a9-8274-bb3cf29a6462 service nova] Releasing lock "refresh_cache-7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1167.538765] env[61868]: DEBUG oslo_vmware.api [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5277061c-dfa6-af4e-a626-d75a55c665a3, 'name': SearchDatastore_Task, 'duration_secs': 0.024249} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.539904] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2ac7d0d-c3c6-47ec-9a19-dedb621c14f9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.547187] env[61868]: DEBUG oslo_vmware.api [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1167.547187] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52873b2e-f364-b481-8fa1-f38b502d8a0f" [ 1167.547187] env[61868]: _type = "Task" [ 1167.547187] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.558745] env[61868]: DEBUG oslo_vmware.api [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52873b2e-f364-b481-8fa1-f38b502d8a0f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.057855] env[61868]: DEBUG oslo_vmware.api [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52873b2e-f364-b481-8fa1-f38b502d8a0f, 'name': SearchDatastore_Task, 'duration_secs': 0.023819} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.058157] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1168.058420] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e/7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1168.058700] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-24d65ac4-ef2c-4efc-aa29-630c31b3f35f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.065034] env[61868]: DEBUG oslo_vmware.api [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1168.065034] env[61868]: value = "task-1316047" [ 1168.065034] env[61868]: _type = "Task" [ 1168.065034] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.073425] env[61868]: DEBUG oslo_vmware.api [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316047, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.574807] env[61868]: DEBUG oslo_vmware.api [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316047, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.076722] env[61868]: DEBUG oslo_vmware.api [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316047, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.577529} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.076971] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e/7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1169.077231] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1169.077497] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a2b06470-cd84-469e-a05b-60eefa2d87e3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.084465] env[61868]: DEBUG oslo_vmware.api [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1169.084465] env[61868]: value = "task-1316048" [ 1169.084465] env[61868]: _type = "Task" [ 1169.084465] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.092690] env[61868]: DEBUG oslo_vmware.api [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316048, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.593791] env[61868]: DEBUG oslo_vmware.api [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316048, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068493} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.594201] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1169.594801] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5145a7c0-e850-4d2c-9f69-8deff22c06bf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.616477] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e/7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1169.616792] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-08e0d439-aa76-4444-8c89-e6396bc1e0e5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.636340] env[61868]: DEBUG oslo_vmware.api [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1169.636340] env[61868]: value = "task-1316049" [ 1169.636340] env[61868]: _type = "Task" [ 1169.636340] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.644332] env[61868]: DEBUG oslo_vmware.api [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316049, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.146126] env[61868]: DEBUG oslo_vmware.api [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316049, 'name': ReconfigVM_Task, 'duration_secs': 0.279363} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.146536] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Reconfigured VM instance instance-0000006b to attach disk [datastore1] 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e/7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1170.147229] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1c34e6da-c95e-4878-8f58-7271b378dac7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.154369] env[61868]: DEBUG oslo_vmware.api [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1170.154369] env[61868]: value = "task-1316050" [ 1170.154369] env[61868]: _type = "Task" [ 1170.154369] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.162109] env[61868]: DEBUG oslo_vmware.api [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316050, 'name': Rename_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.664168] env[61868]: DEBUG oslo_vmware.api [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316050, 'name': Rename_Task, 'duration_secs': 0.134407} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.664514] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1170.664643] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-17fbad87-fa1e-457f-877a-cbee68308261 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.670824] env[61868]: DEBUG oslo_vmware.api [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1170.670824] env[61868]: value = "task-1316051" [ 1170.670824] env[61868]: _type = "Task" [ 1170.670824] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.678356] env[61868]: DEBUG oslo_vmware.api [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316051, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.181039] env[61868]: DEBUG oslo_vmware.api [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316051, 'name': PowerOnVM_Task, 'duration_secs': 0.461143} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.181039] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1171.181235] env[61868]: INFO nova.compute.manager [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Took 7.23 seconds to spawn the instance on the hypervisor. [ 1171.181366] env[61868]: DEBUG nova.compute.manager [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1171.182199] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45c82544-1300-42a9-b393-b3138dd1caa7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.699816] env[61868]: INFO nova.compute.manager [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Took 11.91 seconds to build instance. [ 1172.003510] env[61868]: DEBUG oslo_vmware.rw_handles [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5298676d-a3c0-f2b8-d97e-6a1089bfbc9b/disk-0.vmdk. {{(pid=61868) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1172.004548] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e215ca49-4f48-4369-9652-64c75374ada1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.010896] env[61868]: DEBUG oslo_vmware.rw_handles [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5298676d-a3c0-f2b8-d97e-6a1089bfbc9b/disk-0.vmdk is in state: ready. {{(pid=61868) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1172.011083] env[61868]: ERROR oslo_vmware.rw_handles [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5298676d-a3c0-f2b8-d97e-6a1089bfbc9b/disk-0.vmdk due to incomplete transfer. [ 1172.011288] env[61868]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-fe2adede-3bf9-4131-957b-3f2915784f20 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.018596] env[61868]: DEBUG oslo_vmware.rw_handles [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5298676d-a3c0-f2b8-d97e-6a1089bfbc9b/disk-0.vmdk. {{(pid=61868) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1172.018839] env[61868]: DEBUG nova.virt.vmwareapi.images [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Uploaded image 456023ac-572a-4238-9427-5a28e4383daf to the Glance image server {{(pid=61868) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1172.021493] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Destroying the VM {{(pid=61868) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1172.021829] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-7f74c505-fab1-4e49-86aa-ca52c174c8c1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.030579] env[61868]: DEBUG nova.compute.manager [req-86166fd6-5326-48e5-a752-b1eef5259ebf req-37832a99-37d9-441d-8e2e-4e5ce477ae0e service nova] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Received event network-changed-2bac9dd8-51e8-4aad-bf36-90e045aa9106 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1172.030884] env[61868]: DEBUG nova.compute.manager [req-86166fd6-5326-48e5-a752-b1eef5259ebf req-37832a99-37d9-441d-8e2e-4e5ce477ae0e service nova] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Refreshing instance network info cache due to event network-changed-2bac9dd8-51e8-4aad-bf36-90e045aa9106. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1172.031174] env[61868]: DEBUG oslo_concurrency.lockutils [req-86166fd6-5326-48e5-a752-b1eef5259ebf req-37832a99-37d9-441d-8e2e-4e5ce477ae0e service nova] Acquiring lock "refresh_cache-7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1172.031378] env[61868]: DEBUG oslo_concurrency.lockutils [req-86166fd6-5326-48e5-a752-b1eef5259ebf req-37832a99-37d9-441d-8e2e-4e5ce477ae0e service nova] Acquired lock "refresh_cache-7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1172.031551] env[61868]: DEBUG nova.network.neutron [req-86166fd6-5326-48e5-a752-b1eef5259ebf req-37832a99-37d9-441d-8e2e-4e5ce477ae0e service nova] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Refreshing network info cache for port 2bac9dd8-51e8-4aad-bf36-90e045aa9106 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1172.033322] env[61868]: DEBUG oslo_vmware.api [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1172.033322] env[61868]: value = "task-1316052" [ 1172.033322] env[61868]: _type = "Task" [ 1172.033322] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.041652] env[61868]: DEBUG oslo_vmware.api [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316052, 'name': Destroy_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.202590] env[61868]: DEBUG oslo_concurrency.lockutils [None req-7e08812e-58e9-4b6f-92f2-fc96021afb43 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.423s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1172.546492] env[61868]: DEBUG oslo_vmware.api [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316052, 'name': Destroy_Task, 'duration_secs': 0.334837} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.546775] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Destroyed the VM [ 1172.547029] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Deleting Snapshot of the VM instance {{(pid=61868) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1172.547289] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-9fcdb8fb-551b-47b4-8888-6e5a2c27b09c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.553395] env[61868]: DEBUG oslo_vmware.api [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1172.553395] env[61868]: value = "task-1316053" [ 1172.553395] env[61868]: _type = "Task" [ 1172.553395] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.561277] env[61868]: DEBUG oslo_vmware.api [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316053, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.802638] env[61868]: DEBUG nova.network.neutron [req-86166fd6-5326-48e5-a752-b1eef5259ebf req-37832a99-37d9-441d-8e2e-4e5ce477ae0e service nova] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Updated VIF entry in instance network info cache for port 2bac9dd8-51e8-4aad-bf36-90e045aa9106. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1172.803203] env[61868]: DEBUG nova.network.neutron [req-86166fd6-5326-48e5-a752-b1eef5259ebf req-37832a99-37d9-441d-8e2e-4e5ce477ae0e service nova] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Updating instance_info_cache with network_info: [{"id": "2bac9dd8-51e8-4aad-bf36-90e045aa9106", "address": "fa:16:3e:d7:73:72", "network": {"id": "16b3638c-1129-49b8-9f22-636ff7db7386", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1544179589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bdab2203834c4ee5bab11aca70e48cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d298db54-f13d-4bf6-b6c2-755074b3047f", "external-id": "nsx-vlan-transportzone-631", "segmentation_id": 631, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2bac9dd8-51", "ovs_interfaceid": "2bac9dd8-51e8-4aad-bf36-90e045aa9106", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1173.062708] env[61868]: DEBUG oslo_vmware.api [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316053, 'name': RemoveSnapshot_Task, 'duration_secs': 0.49013} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.062938] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Deleted Snapshot of the VM instance {{(pid=61868) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1173.063183] env[61868]: INFO nova.compute.manager [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Took 12.62 seconds to snapshot the instance on the hypervisor. [ 1173.306465] env[61868]: DEBUG oslo_concurrency.lockutils [req-86166fd6-5326-48e5-a752-b1eef5259ebf req-37832a99-37d9-441d-8e2e-4e5ce477ae0e service nova] Releasing lock "refresh_cache-7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1173.604788] env[61868]: DEBUG nova.compute.manager [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Found 3 images (rotation: 2) {{(pid=61868) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 1173.605000] env[61868]: DEBUG nova.compute.manager [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Rotating out 1 backups {{(pid=61868) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4562}} [ 1173.605212] env[61868]: DEBUG nova.compute.manager [None req-fc05fb0f-b70e-422c-bc23-c8b0abfbf5ae tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Deleting image e70ee3c8-6ebd-4457-8ba6-ef94cd5a1e56 {{(pid=61868) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4567}} [ 1175.972036] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b6079163-91eb-414a-a0d7-2fbe0a1b7ab1 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "f2d864f1-cd11-4b62-857d-789cf045f22c" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1175.972399] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b6079163-91eb-414a-a0d7-2fbe0a1b7ab1 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "f2d864f1-cd11-4b62-857d-789cf045f22c" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1175.972399] env[61868]: DEBUG nova.compute.manager [None req-b6079163-91eb-414a-a0d7-2fbe0a1b7ab1 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1175.973405] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48105228-22eb-4671-9cca-cda5a84aba68 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.980046] env[61868]: DEBUG nova.compute.manager [None req-b6079163-91eb-414a-a0d7-2fbe0a1b7ab1 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61868) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1175.980626] env[61868]: DEBUG nova.objects.instance [None req-b6079163-91eb-414a-a0d7-2fbe0a1b7ab1 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lazy-loading 'flavor' on Instance uuid f2d864f1-cd11-4b62-857d-789cf045f22c {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1176.485965] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6079163-91eb-414a-a0d7-2fbe0a1b7ab1 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1176.486319] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-14d66495-e48f-477c-8e25-e38caac9f2bd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.493857] env[61868]: DEBUG oslo_vmware.api [None req-b6079163-91eb-414a-a0d7-2fbe0a1b7ab1 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1176.493857] env[61868]: value = "task-1316054" [ 1176.493857] env[61868]: _type = "Task" [ 1176.493857] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.502742] env[61868]: DEBUG oslo_vmware.api [None req-b6079163-91eb-414a-a0d7-2fbe0a1b7ab1 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316054, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.560781] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5227b7d7-2443-45fe-8987-168870d9c1a9 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "afd7ff53-f164-4f3d-9ece-28a10245e38c" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1176.561030] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5227b7d7-2443-45fe-8987-168870d9c1a9 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "afd7ff53-f164-4f3d-9ece-28a10245e38c" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1177.005230] env[61868]: DEBUG oslo_vmware.api [None req-b6079163-91eb-414a-a0d7-2fbe0a1b7ab1 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316054, 'name': PowerOffVM_Task, 'duration_secs': 0.169846} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.005596] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6079163-91eb-414a-a0d7-2fbe0a1b7ab1 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1177.005695] env[61868]: DEBUG nova.compute.manager [None req-b6079163-91eb-414a-a0d7-2fbe0a1b7ab1 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1177.006405] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8512ceab-7e1b-4a6f-9622-aa1fb4f6551d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.063954] env[61868]: DEBUG nova.compute.utils [None req-5227b7d7-2443-45fe-8987-168870d9c1a9 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1177.517444] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b6079163-91eb-414a-a0d7-2fbe0a1b7ab1 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "f2d864f1-cd11-4b62-857d-789cf045f22c" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.545s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1177.566533] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5227b7d7-2443-45fe-8987-168870d9c1a9 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "afd7ff53-f164-4f3d-9ece-28a10245e38c" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.005s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1178.626058] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5227b7d7-2443-45fe-8987-168870d9c1a9 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "afd7ff53-f164-4f3d-9ece-28a10245e38c" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1178.626058] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5227b7d7-2443-45fe-8987-168870d9c1a9 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "afd7ff53-f164-4f3d-9ece-28a10245e38c" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1178.626058] env[61868]: INFO nova.compute.manager [None req-5227b7d7-2443-45fe-8987-168870d9c1a9 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Attaching volume c692c9a5-d270-4a97-926a-debe09de58c8 to /dev/sdb [ 1178.657567] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49e710e7-1876-4a9a-997b-ccd14ee848fd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.664386] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8c812b8-90aa-493d-a028-c3e059cd3c2a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.677414] env[61868]: DEBUG nova.virt.block_device [None req-5227b7d7-2443-45fe-8987-168870d9c1a9 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Updating existing volume attachment record: f2d69156-06fa-4f5f-812a-0fac23cb3f57 {{(pid=61868) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1179.049186] env[61868]: DEBUG nova.compute.manager [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Stashing vm_state: stopped {{(pid=61868) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1179.569479] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1179.569761] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1180.074688] env[61868]: INFO nova.compute.claims [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1180.581208] env[61868]: INFO nova.compute.resource_tracker [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Updating resource usage from migration 48264e8c-0b0d-4329-9a88-9433a33519bc [ 1180.650465] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2de44025-ba80-4379-8f08-732fe2aeabbb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.657626] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01df89df-5be1-4162-bf2d-7278adbb002d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.687379] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e5ccb03-d911-4ae2-a5d5-0d6fb4c28001 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.693980] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2bc35d3-7c0a-4816-9032-c3fddf197cca {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.706305] env[61868]: DEBUG nova.compute.provider_tree [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1181.209433] env[61868]: DEBUG nova.scheduler.client.report [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1181.714367] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.144s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1181.714578] env[61868]: INFO nova.compute.manager [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Migrating [ 1182.229854] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "refresh_cache-f2d864f1-cd11-4b62-857d-789cf045f22c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1182.230301] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquired lock "refresh_cache-f2d864f1-cd11-4b62-857d-789cf045f22c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1182.230301] env[61868]: DEBUG nova.network.neutron [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1182.937586] env[61868]: DEBUG nova.network.neutron [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Updating instance_info_cache with network_info: [{"id": "4c07a7f4-8d47-494b-977a-1aa62477db98", "address": "fa:16:3e:b7:7b:cc", "network": {"id": "7e8cbecf-dfdd-43f2-8745-ae1d2d55d309", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-33935587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a7014141b24101ba7920bc83be4a9e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4db2961d-273d-4634-9d06-a94fa9d384fb", "external-id": "nsx-vlan-transportzone-572", "segmentation_id": 572, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c07a7f4-8d", "ovs_interfaceid": "4c07a7f4-8d47-494b-977a-1aa62477db98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1183.219898] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-5227b7d7-2443-45fe-8987-168870d9c1a9 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Volume attach. Driver type: vmdk {{(pid=61868) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1183.220201] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-5227b7d7-2443-45fe-8987-168870d9c1a9 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281638', 'volume_id': 'c692c9a5-d270-4a97-926a-debe09de58c8', 'name': 'volume-c692c9a5-d270-4a97-926a-debe09de58c8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'afd7ff53-f164-4f3d-9ece-28a10245e38c', 'attached_at': '', 'detached_at': '', 'volume_id': 'c692c9a5-d270-4a97-926a-debe09de58c8', 'serial': 'c692c9a5-d270-4a97-926a-debe09de58c8'} {{(pid=61868) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1183.221086] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b61bf5f-113d-4859-b28b-0e7d3368826d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.237152] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-638e30e8-aaf4-4711-a1e8-e7e713619832 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.260221] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-5227b7d7-2443-45fe-8987-168870d9c1a9 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] volume-c692c9a5-d270-4a97-926a-debe09de58c8/volume-c692c9a5-d270-4a97-926a-debe09de58c8.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1183.260460] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d11da33b-b4ee-4eb5-83e0-652f5196263f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.277956] env[61868]: DEBUG oslo_vmware.api [None req-5227b7d7-2443-45fe-8987-168870d9c1a9 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 1183.277956] env[61868]: value = "task-1316057" [ 1183.277956] env[61868]: _type = "Task" [ 1183.277956] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.285375] env[61868]: DEBUG oslo_vmware.api [None req-5227b7d7-2443-45fe-8987-168870d9c1a9 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316057, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.440520] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Releasing lock "refresh_cache-f2d864f1-cd11-4b62-857d-789cf045f22c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1183.788102] env[61868]: DEBUG oslo_vmware.api [None req-5227b7d7-2443-45fe-8987-168870d9c1a9 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316057, 'name': ReconfigVM_Task, 'duration_secs': 0.329604} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.788361] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-5227b7d7-2443-45fe-8987-168870d9c1a9 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Reconfigured VM instance instance-0000006a to attach disk [datastore1] volume-c692c9a5-d270-4a97-926a-debe09de58c8/volume-c692c9a5-d270-4a97-926a-debe09de58c8.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1183.793150] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7e73b748-7b9b-4fcd-b480-d80d0c556ae6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.807193] env[61868]: DEBUG oslo_vmware.api [None req-5227b7d7-2443-45fe-8987-168870d9c1a9 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 1183.807193] env[61868]: value = "task-1316058" [ 1183.807193] env[61868]: _type = "Task" [ 1183.807193] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.814206] env[61868]: DEBUG oslo_vmware.api [None req-5227b7d7-2443-45fe-8987-168870d9c1a9 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316058, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.316807] env[61868]: DEBUG oslo_vmware.api [None req-5227b7d7-2443-45fe-8987-168870d9c1a9 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316058, 'name': ReconfigVM_Task, 'duration_secs': 0.125376} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.317128] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-5227b7d7-2443-45fe-8987-168870d9c1a9 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281638', 'volume_id': 'c692c9a5-d270-4a97-926a-debe09de58c8', 'name': 'volume-c692c9a5-d270-4a97-926a-debe09de58c8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'afd7ff53-f164-4f3d-9ece-28a10245e38c', 'attached_at': '', 'detached_at': '', 'volume_id': 'c692c9a5-d270-4a97-926a-debe09de58c8', 'serial': 'c692c9a5-d270-4a97-926a-debe09de58c8'} {{(pid=61868) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1184.956728] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a79c922-d1a0-43f3-8c3c-6db8bdf90676 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.976064] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Updating instance 'f2d864f1-cd11-4b62-857d-789cf045f22c' progress to 0 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1185.359687] env[61868]: DEBUG nova.objects.instance [None req-5227b7d7-2443-45fe-8987-168870d9c1a9 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lazy-loading 'flavor' on Instance uuid afd7ff53-f164-4f3d-9ece-28a10245e38c {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1185.482055] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1185.482055] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5e066411-76b9-42ea-9f8e-822d599a6204 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.489280] env[61868]: DEBUG oslo_vmware.api [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1185.489280] env[61868]: value = "task-1316059" [ 1185.489280] env[61868]: _type = "Task" [ 1185.489280] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.497985] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] VM already powered off {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1185.498256] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Updating instance 'f2d864f1-cd11-4b62-857d-789cf045f22c' progress to 17 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1185.867065] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5227b7d7-2443-45fe-8987-168870d9c1a9 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "afd7ff53-f164-4f3d-9ece-28a10245e38c" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.241s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1186.004042] env[61868]: DEBUG nova.virt.hardware [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1186.004318] env[61868]: DEBUG nova.virt.hardware [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1186.004483] env[61868]: DEBUG nova.virt.hardware [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1186.004671] env[61868]: DEBUG nova.virt.hardware [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1186.004822] env[61868]: DEBUG nova.virt.hardware [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1186.004979] env[61868]: DEBUG nova.virt.hardware [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1186.005244] env[61868]: DEBUG nova.virt.hardware [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1186.005421] env[61868]: DEBUG nova.virt.hardware [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1186.005590] env[61868]: DEBUG nova.virt.hardware [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1186.005770] env[61868]: DEBUG nova.virt.hardware [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1186.005949] env[61868]: DEBUG nova.virt.hardware [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1186.011354] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-16eebbdf-b347-4706-87ee-d81975593e14 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.027376] env[61868]: DEBUG oslo_vmware.api [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1186.027376] env[61868]: value = "task-1316060" [ 1186.027376] env[61868]: _type = "Task" [ 1186.027376] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.036153] env[61868]: DEBUG oslo_concurrency.lockutils [None req-270ceca5-8150-4349-8a2d-d9917951e274 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "afd7ff53-f164-4f3d-9ece-28a10245e38c" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1186.036399] env[61868]: DEBUG oslo_concurrency.lockutils [None req-270ceca5-8150-4349-8a2d-d9917951e274 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "afd7ff53-f164-4f3d-9ece-28a10245e38c" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1186.037763] env[61868]: DEBUG oslo_vmware.api [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316060, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.537302] env[61868]: DEBUG oslo_vmware.api [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316060, 'name': ReconfigVM_Task, 'duration_secs': 0.149334} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.537774] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Updating instance 'f2d864f1-cd11-4b62-857d-789cf045f22c' progress to 33 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1186.542788] env[61868]: INFO nova.compute.manager [None req-270ceca5-8150-4349-8a2d-d9917951e274 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Detaching volume c692c9a5-d270-4a97-926a-debe09de58c8 [ 1186.585471] env[61868]: INFO nova.virt.block_device [None req-270ceca5-8150-4349-8a2d-d9917951e274 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Attempting to driver detach volume c692c9a5-d270-4a97-926a-debe09de58c8 from mountpoint /dev/sdb [ 1186.585723] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-270ceca5-8150-4349-8a2d-d9917951e274 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Volume detach. Driver type: vmdk {{(pid=61868) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1186.585911] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-270ceca5-8150-4349-8a2d-d9917951e274 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281638', 'volume_id': 'c692c9a5-d270-4a97-926a-debe09de58c8', 'name': 'volume-c692c9a5-d270-4a97-926a-debe09de58c8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'afd7ff53-f164-4f3d-9ece-28a10245e38c', 'attached_at': '', 'detached_at': '', 'volume_id': 'c692c9a5-d270-4a97-926a-debe09de58c8', 'serial': 'c692c9a5-d270-4a97-926a-debe09de58c8'} {{(pid=61868) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1186.587049] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac997e6d-fc34-4efc-add5-d0ed304dd7dc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.611112] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0181223d-f48a-4c9c-af36-b84c2ad88cf1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.617861] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4540806-291a-45b7-8135-bec07c385244 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.637589] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ebcda50-063d-4795-a1ce-100c78f223f5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.651956] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-270ceca5-8150-4349-8a2d-d9917951e274 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] The volume has not been displaced from its original location: [datastore1] volume-c692c9a5-d270-4a97-926a-debe09de58c8/volume-c692c9a5-d270-4a97-926a-debe09de58c8.vmdk. No consolidation needed. {{(pid=61868) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1186.657295] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-270ceca5-8150-4349-8a2d-d9917951e274 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Reconfiguring VM instance instance-0000006a to detach disk 2001 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1186.657599] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aed9aa38-e9f7-40ab-bb7c-6967aab8f52c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.674414] env[61868]: DEBUG oslo_vmware.api [None req-270ceca5-8150-4349-8a2d-d9917951e274 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 1186.674414] env[61868]: value = "task-1316061" [ 1186.674414] env[61868]: _type = "Task" [ 1186.674414] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.681567] env[61868]: DEBUG oslo_vmware.api [None req-270ceca5-8150-4349-8a2d-d9917951e274 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316061, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.044252] env[61868]: DEBUG nova.virt.hardware [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1187.044503] env[61868]: DEBUG nova.virt.hardware [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1187.044646] env[61868]: DEBUG nova.virt.hardware [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1187.044834] env[61868]: DEBUG nova.virt.hardware [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1187.044984] env[61868]: DEBUG nova.virt.hardware [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1187.045169] env[61868]: DEBUG nova.virt.hardware [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1187.045394] env[61868]: DEBUG nova.virt.hardware [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1187.045564] env[61868]: DEBUG nova.virt.hardware [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1187.045735] env[61868]: DEBUG nova.virt.hardware [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1187.045901] env[61868]: DEBUG nova.virt.hardware [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1187.046098] env[61868]: DEBUG nova.virt.hardware [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1187.051343] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Reconfiguring VM instance instance-00000068 to detach disk 2000 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1187.051628] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6b6c387a-0ee4-4791-a56f-b95fdf59e4e1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.070068] env[61868]: DEBUG oslo_vmware.api [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1187.070068] env[61868]: value = "task-1316062" [ 1187.070068] env[61868]: _type = "Task" [ 1187.070068] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.080112] env[61868]: DEBUG oslo_vmware.api [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316062, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.183240] env[61868]: DEBUG oslo_vmware.api [None req-270ceca5-8150-4349-8a2d-d9917951e274 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316061, 'name': ReconfigVM_Task, 'duration_secs': 0.198274} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.183529] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-270ceca5-8150-4349-8a2d-d9917951e274 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Reconfigured VM instance instance-0000006a to detach disk 2001 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1187.188157] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-66d2a5ba-277a-41ea-b928-88b263aa1e53 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.202366] env[61868]: DEBUG oslo_vmware.api [None req-270ceca5-8150-4349-8a2d-d9917951e274 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 1187.202366] env[61868]: value = "task-1316063" [ 1187.202366] env[61868]: _type = "Task" [ 1187.202366] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.209341] env[61868]: DEBUG oslo_vmware.api [None req-270ceca5-8150-4349-8a2d-d9917951e274 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316063, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.579789] env[61868]: DEBUG oslo_vmware.api [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316062, 'name': ReconfigVM_Task, 'duration_secs': 0.142387} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.580165] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Reconfigured VM instance instance-00000068 to detach disk 2000 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1187.580825] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bcbd720-b421-4066-9b2b-dbf16ab37a11 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.601878] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] f2d864f1-cd11-4b62-857d-789cf045f22c/f2d864f1-cd11-4b62-857d-789cf045f22c.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1187.602132] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8f726cde-b468-4710-a85c-05bf6f9ad03a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.619292] env[61868]: DEBUG oslo_vmware.api [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1187.619292] env[61868]: value = "task-1316064" [ 1187.619292] env[61868]: _type = "Task" [ 1187.619292] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.628060] env[61868]: DEBUG oslo_vmware.api [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316064, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.711678] env[61868]: DEBUG oslo_vmware.api [None req-270ceca5-8150-4349-8a2d-d9917951e274 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316063, 'name': ReconfigVM_Task, 'duration_secs': 0.124706} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.712037] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-270ceca5-8150-4349-8a2d-d9917951e274 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281638', 'volume_id': 'c692c9a5-d270-4a97-926a-debe09de58c8', 'name': 'volume-c692c9a5-d270-4a97-926a-debe09de58c8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'afd7ff53-f164-4f3d-9ece-28a10245e38c', 'attached_at': '', 'detached_at': '', 'volume_id': 'c692c9a5-d270-4a97-926a-debe09de58c8', 'serial': 'c692c9a5-d270-4a97-926a-debe09de58c8'} {{(pid=61868) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1188.128867] env[61868]: DEBUG oslo_vmware.api [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316064, 'name': ReconfigVM_Task, 'duration_secs': 0.236929} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.129166] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Reconfigured VM instance instance-00000068 to attach disk [datastore2] f2d864f1-cd11-4b62-857d-789cf045f22c/f2d864f1-cd11-4b62-857d-789cf045f22c.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1188.129433] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Updating instance 'f2d864f1-cd11-4b62-857d-789cf045f22c' progress to 50 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1188.252504] env[61868]: DEBUG nova.objects.instance [None req-270ceca5-8150-4349-8a2d-d9917951e274 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lazy-loading 'flavor' on Instance uuid afd7ff53-f164-4f3d-9ece-28a10245e38c {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1188.636156] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81118b1c-2365-441c-88f1-5ca8aa0e698c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.654729] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-995c40a4-81c9-4047-8d7b-de7f7ce55a3e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.671496] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Updating instance 'f2d864f1-cd11-4b62-857d-789cf045f22c' progress to 67 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1189.210087] env[61868]: DEBUG nova.network.neutron [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Port 4c07a7f4-8d47-494b-977a-1aa62477db98 binding to destination host cpu-1 is already ACTIVE {{(pid=61868) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1189.257954] env[61868]: DEBUG oslo_concurrency.lockutils [None req-270ceca5-8150-4349-8a2d-d9917951e274 tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "afd7ff53-f164-4f3d-9ece-28a10245e38c" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.221s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1190.232391] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "f2d864f1-cd11-4b62-857d-789cf045f22c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1190.232720] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "f2d864f1-cd11-4b62-857d-789cf045f22c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1190.232856] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "f2d864f1-cd11-4b62-857d-789cf045f22c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1190.290220] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b550cd86-c436-4d4b-a8b0-b5a7a7caacbe tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "afd7ff53-f164-4f3d-9ece-28a10245e38c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1190.290488] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b550cd86-c436-4d4b-a8b0-b5a7a7caacbe tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "afd7ff53-f164-4f3d-9ece-28a10245e38c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1190.290695] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b550cd86-c436-4d4b-a8b0-b5a7a7caacbe tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "afd7ff53-f164-4f3d-9ece-28a10245e38c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1190.290886] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b550cd86-c436-4d4b-a8b0-b5a7a7caacbe tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "afd7ff53-f164-4f3d-9ece-28a10245e38c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1190.291108] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b550cd86-c436-4d4b-a8b0-b5a7a7caacbe tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "afd7ff53-f164-4f3d-9ece-28a10245e38c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1190.292994] env[61868]: INFO nova.compute.manager [None req-b550cd86-c436-4d4b-a8b0-b5a7a7caacbe tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Terminating instance [ 1190.294742] env[61868]: DEBUG nova.compute.manager [None req-b550cd86-c436-4d4b-a8b0-b5a7a7caacbe tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1190.294931] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-b550cd86-c436-4d4b-a8b0-b5a7a7caacbe tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1190.295825] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa776ef1-040b-4b46-bbfd-60416d2cfd39 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.303672] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-b550cd86-c436-4d4b-a8b0-b5a7a7caacbe tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1190.303901] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7f5f970d-9a02-455a-8e3e-78235c3773cc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.310328] env[61868]: DEBUG oslo_vmware.api [None req-b550cd86-c436-4d4b-a8b0-b5a7a7caacbe tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 1190.310328] env[61868]: value = "task-1316065" [ 1190.310328] env[61868]: _type = "Task" [ 1190.310328] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.317550] env[61868]: DEBUG oslo_vmware.api [None req-b550cd86-c436-4d4b-a8b0-b5a7a7caacbe tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316065, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.820277] env[61868]: DEBUG oslo_vmware.api [None req-b550cd86-c436-4d4b-a8b0-b5a7a7caacbe tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316065, 'name': PowerOffVM_Task, 'duration_secs': 0.143578} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.820533] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-b550cd86-c436-4d4b-a8b0-b5a7a7caacbe tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1190.820700] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-b550cd86-c436-4d4b-a8b0-b5a7a7caacbe tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1190.820942] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-26f3e626-2467-425a-af73-1fd8c986546a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.882423] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-b550cd86-c436-4d4b-a8b0-b5a7a7caacbe tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1190.882655] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-b550cd86-c436-4d4b-a8b0-b5a7a7caacbe tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1190.882841] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-b550cd86-c436-4d4b-a8b0-b5a7a7caacbe tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Deleting the datastore file [datastore1] afd7ff53-f164-4f3d-9ece-28a10245e38c {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1190.883148] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2f2d814f-8c8b-4d75-8762-9e8d9c99d08e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.889820] env[61868]: DEBUG oslo_vmware.api [None req-b550cd86-c436-4d4b-a8b0-b5a7a7caacbe tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for the task: (returnval){ [ 1190.889820] env[61868]: value = "task-1316067" [ 1190.889820] env[61868]: _type = "Task" [ 1190.889820] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.897052] env[61868]: DEBUG oslo_vmware.api [None req-b550cd86-c436-4d4b-a8b0-b5a7a7caacbe tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316067, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.283203] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "refresh_cache-f2d864f1-cd11-4b62-857d-789cf045f22c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1191.283468] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquired lock "refresh_cache-f2d864f1-cd11-4b62-857d-789cf045f22c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1191.283685] env[61868]: DEBUG nova.network.neutron [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1191.399773] env[61868]: DEBUG oslo_vmware.api [None req-b550cd86-c436-4d4b-a8b0-b5a7a7caacbe tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Task: {'id': task-1316067, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.126981} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.400048] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-b550cd86-c436-4d4b-a8b0-b5a7a7caacbe tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1191.400240] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-b550cd86-c436-4d4b-a8b0-b5a7a7caacbe tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1191.400423] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-b550cd86-c436-4d4b-a8b0-b5a7a7caacbe tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1191.400596] env[61868]: INFO nova.compute.manager [None req-b550cd86-c436-4d4b-a8b0-b5a7a7caacbe tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1191.400841] env[61868]: DEBUG oslo.service.loopingcall [None req-b550cd86-c436-4d4b-a8b0-b5a7a7caacbe tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1191.401050] env[61868]: DEBUG nova.compute.manager [-] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1191.401147] env[61868]: DEBUG nova.network.neutron [-] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1191.998397] env[61868]: DEBUG nova.network.neutron [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Updating instance_info_cache with network_info: [{"id": "4c07a7f4-8d47-494b-977a-1aa62477db98", "address": "fa:16:3e:b7:7b:cc", "network": {"id": "7e8cbecf-dfdd-43f2-8745-ae1d2d55d309", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-33935587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a7014141b24101ba7920bc83be4a9e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4db2961d-273d-4634-9d06-a94fa9d384fb", "external-id": "nsx-vlan-transportzone-572", "segmentation_id": 572, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c07a7f4-8d", "ovs_interfaceid": "4c07a7f4-8d47-494b-977a-1aa62477db98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1192.060081] env[61868]: DEBUG nova.compute.manager [req-6d4f04cd-1d6c-4f5b-93d4-fe997b715130 req-4b1ab211-4dc1-46a5-b05f-60f72411a969 service nova] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Received event network-vif-deleted-4f02e3ef-73cc-42e9-92a7-4366bb33e757 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1192.060311] env[61868]: INFO nova.compute.manager [req-6d4f04cd-1d6c-4f5b-93d4-fe997b715130 req-4b1ab211-4dc1-46a5-b05f-60f72411a969 service nova] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Neutron deleted interface 4f02e3ef-73cc-42e9-92a7-4366bb33e757; detaching it from the instance and deleting it from the info cache [ 1192.060542] env[61868]: DEBUG nova.network.neutron [req-6d4f04cd-1d6c-4f5b-93d4-fe997b715130 req-4b1ab211-4dc1-46a5-b05f-60f72411a969 service nova] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1192.501076] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Releasing lock "refresh_cache-f2d864f1-cd11-4b62-857d-789cf045f22c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1192.534869] env[61868]: DEBUG nova.network.neutron [-] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1192.564738] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fd26965f-5c13-4173-a67c-99341afc7bd7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.573667] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fe0602f-e9dc-4d4f-b4c1-d482fbebb5eb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.597430] env[61868]: DEBUG nova.compute.manager [req-6d4f04cd-1d6c-4f5b-93d4-fe997b715130 req-4b1ab211-4dc1-46a5-b05f-60f72411a969 service nova] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Detach interface failed, port_id=4f02e3ef-73cc-42e9-92a7-4366bb33e757, reason: Instance afd7ff53-f164-4f3d-9ece-28a10245e38c could not be found. {{(pid=61868) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1193.025855] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e834b20-0ffc-4da8-b4ad-9c436d5c157a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.043667] env[61868]: INFO nova.compute.manager [-] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Took 1.64 seconds to deallocate network for instance. [ 1193.045885] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfe252c8-e5e8-403b-bc65-e7b1d7cd1e4f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.054969] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Updating instance 'f2d864f1-cd11-4b62-857d-789cf045f22c' progress to 83 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1193.552888] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b550cd86-c436-4d4b-a8b0-b5a7a7caacbe tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1193.553258] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b550cd86-c436-4d4b-a8b0-b5a7a7caacbe tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1193.553439] env[61868]: DEBUG nova.objects.instance [None req-b550cd86-c436-4d4b-a8b0-b5a7a7caacbe tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lazy-loading 'resources' on Instance uuid afd7ff53-f164-4f3d-9ece-28a10245e38c {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1193.560666] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ab052837-f3b4-42fb-a690-a7cc86a20de0 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Updating instance 'f2d864f1-cd11-4b62-857d-789cf045f22c' progress to 100 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1194.130149] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8cebbf4-9ad4-47e4-a64f-bcc8c07dcd5a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.137739] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fe3f59c-84a9-475f-aaf0-a7fa13183198 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.168395] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad7227a5-e64a-47ab-9065-0307c6671116 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.177726] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b461992b-5c5b-49a2-8a7c-a0d5a9ea48d5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.190470] env[61868]: DEBUG nova.compute.provider_tree [None req-b550cd86-c436-4d4b-a8b0-b5a7a7caacbe tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1194.694012] env[61868]: DEBUG nova.scheduler.client.report [None req-b550cd86-c436-4d4b-a8b0-b5a7a7caacbe tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1195.198520] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b550cd86-c436-4d4b-a8b0-b5a7a7caacbe tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.645s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1195.220315] env[61868]: INFO nova.scheduler.client.report [None req-b550cd86-c436-4d4b-a8b0-b5a7a7caacbe tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Deleted allocations for instance afd7ff53-f164-4f3d-9ece-28a10245e38c [ 1195.727056] env[61868]: DEBUG oslo_concurrency.lockutils [None req-b550cd86-c436-4d4b-a8b0-b5a7a7caacbe tempest-AttachVolumeNegativeTest-2120203638 tempest-AttachVolumeNegativeTest-2120203638-project-member] Lock "afd7ff53-f164-4f3d-9ece-28a10245e38c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.436s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1196.167034] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c7288a36-bde2-4341-bda0-07af3611a0f4 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "f2d864f1-cd11-4b62-857d-789cf045f22c" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1196.167350] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c7288a36-bde2-4341-bda0-07af3611a0f4 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "f2d864f1-cd11-4b62-857d-789cf045f22c" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1196.167553] env[61868]: DEBUG nova.compute.manager [None req-c7288a36-bde2-4341-bda0-07af3611a0f4 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Going to confirm migration 4 {{(pid=61868) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1196.753964] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c7288a36-bde2-4341-bda0-07af3611a0f4 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "refresh_cache-f2d864f1-cd11-4b62-857d-789cf045f22c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1196.754236] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c7288a36-bde2-4341-bda0-07af3611a0f4 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquired lock "refresh_cache-f2d864f1-cd11-4b62-857d-789cf045f22c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1196.754360] env[61868]: DEBUG nova.network.neutron [None req-c7288a36-bde2-4341-bda0-07af3611a0f4 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1196.754550] env[61868]: DEBUG nova.objects.instance [None req-c7288a36-bde2-4341-bda0-07af3611a0f4 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lazy-loading 'info_cache' on Instance uuid f2d864f1-cd11-4b62-857d-789cf045f22c {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1197.971385] env[61868]: DEBUG nova.network.neutron [None req-c7288a36-bde2-4341-bda0-07af3611a0f4 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Updating instance_info_cache with network_info: [{"id": "4c07a7f4-8d47-494b-977a-1aa62477db98", "address": "fa:16:3e:b7:7b:cc", "network": {"id": "7e8cbecf-dfdd-43f2-8745-ae1d2d55d309", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-33935587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a7014141b24101ba7920bc83be4a9e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4db2961d-273d-4634-9d06-a94fa9d384fb", "external-id": "nsx-vlan-transportzone-572", "segmentation_id": 572, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c07a7f4-8d", "ovs_interfaceid": "4c07a7f4-8d47-494b-977a-1aa62477db98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1198.474107] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c7288a36-bde2-4341-bda0-07af3611a0f4 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Releasing lock "refresh_cache-f2d864f1-cd11-4b62-857d-789cf045f22c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1198.474409] env[61868]: DEBUG nova.objects.instance [None req-c7288a36-bde2-4341-bda0-07af3611a0f4 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lazy-loading 'migration_context' on Instance uuid f2d864f1-cd11-4b62-857d-789cf045f22c {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1198.978082] env[61868]: DEBUG nova.objects.base [None req-c7288a36-bde2-4341-bda0-07af3611a0f4 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61868) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1198.978565] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c566a3f-1587-40cf-a3fe-942633a0ba24 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.997352] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f74d968-da84-4a51-9677-716a5125661f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.002778] env[61868]: DEBUG oslo_vmware.api [None req-c7288a36-bde2-4341-bda0-07af3611a0f4 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1199.002778] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]521ab1aa-8435-3e2e-5710-de31263ad035" [ 1199.002778] env[61868]: _type = "Task" [ 1199.002778] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.010248] env[61868]: DEBUG oslo_vmware.api [None req-c7288a36-bde2-4341-bda0-07af3611a0f4 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]521ab1aa-8435-3e2e-5710-de31263ad035, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.512721] env[61868]: DEBUG oslo_vmware.api [None req-c7288a36-bde2-4341-bda0-07af3611a0f4 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]521ab1aa-8435-3e2e-5710-de31263ad035, 'name': SearchDatastore_Task, 'duration_secs': 0.00905} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1199.513027] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c7288a36-bde2-4341-bda0-07af3611a0f4 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1199.513263] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c7288a36-bde2-4341-bda0-07af3611a0f4 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1200.076225] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b857eaf-31db-44fe-998d-0768c7f9b3ba {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.084081] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2579950-f189-41f7-aaca-42a69e0ce475 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.114480] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d89fab26-a457-4305-9a6b-a2ebeb8e6a07 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.121586] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7190ab84-8b41-4a11-8bb9-137a46d384b1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.134343] env[61868]: DEBUG nova.compute.provider_tree [None req-c7288a36-bde2-4341-bda0-07af3611a0f4 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1200.637492] env[61868]: DEBUG nova.scheduler.client.report [None req-c7288a36-bde2-4341-bda0-07af3611a0f4 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1201.647709] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c7288a36-bde2-4341-bda0-07af3611a0f4 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.134s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1201.648092] env[61868]: DEBUG nova.compute.manager [None req-c7288a36-bde2-4341-bda0-07af3611a0f4 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=61868) _confirm_resize /opt/stack/nova/nova/compute/manager.py:4909}} [ 1202.201192] env[61868]: INFO nova.scheduler.client.report [None req-c7288a36-bde2-4341-bda0-07af3611a0f4 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Deleted allocation for migration 48264e8c-0b0d-4329-9a88-9433a33519bc [ 1202.706685] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c7288a36-bde2-4341-bda0-07af3611a0f4 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "f2d864f1-cd11-4b62-857d-789cf045f22c" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.539s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1203.070960] env[61868]: DEBUG nova.objects.instance [None req-8e87ffa1-9606-42a3-87ca-4256eca2eb15 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lazy-loading 'flavor' on Instance uuid f2d864f1-cd11-4b62-857d-789cf045f22c {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1203.575834] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8e87ffa1-9606-42a3-87ca-4256eca2eb15 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "refresh_cache-f2d864f1-cd11-4b62-857d-789cf045f22c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1203.576056] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8e87ffa1-9606-42a3-87ca-4256eca2eb15 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquired lock "refresh_cache-f2d864f1-cd11-4b62-857d-789cf045f22c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1203.576167] env[61868]: DEBUG nova.network.neutron [None req-8e87ffa1-9606-42a3-87ca-4256eca2eb15 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1203.576361] env[61868]: DEBUG nova.objects.instance [None req-8e87ffa1-9606-42a3-87ca-4256eca2eb15 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lazy-loading 'info_cache' on Instance uuid f2d864f1-cd11-4b62-857d-789cf045f22c {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1204.080626] env[61868]: DEBUG nova.objects.base [None req-8e87ffa1-9606-42a3-87ca-4256eca2eb15 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=61868) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1204.781155] env[61868]: DEBUG nova.network.neutron [None req-8e87ffa1-9606-42a3-87ca-4256eca2eb15 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Updating instance_info_cache with network_info: [{"id": "4c07a7f4-8d47-494b-977a-1aa62477db98", "address": "fa:16:3e:b7:7b:cc", "network": {"id": "7e8cbecf-dfdd-43f2-8745-ae1d2d55d309", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-33935587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a7014141b24101ba7920bc83be4a9e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4db2961d-273d-4634-9d06-a94fa9d384fb", "external-id": "nsx-vlan-transportzone-572", "segmentation_id": 572, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c07a7f4-8d", "ovs_interfaceid": "4c07a7f4-8d47-494b-977a-1aa62477db98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1205.285057] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8e87ffa1-9606-42a3-87ca-4256eca2eb15 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Releasing lock "refresh_cache-f2d864f1-cd11-4b62-857d-789cf045f22c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1205.788307] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e87ffa1-9606-42a3-87ca-4256eca2eb15 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1205.788628] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-40de93f9-55cb-49ed-a351-65fa6b346520 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.796195] env[61868]: DEBUG oslo_vmware.api [None req-8e87ffa1-9606-42a3-87ca-4256eca2eb15 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1205.796195] env[61868]: value = "task-1316071" [ 1205.796195] env[61868]: _type = "Task" [ 1205.796195] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.803551] env[61868]: DEBUG oslo_vmware.api [None req-8e87ffa1-9606-42a3-87ca-4256eca2eb15 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316071, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.306153] env[61868]: DEBUG oslo_vmware.api [None req-8e87ffa1-9606-42a3-87ca-4256eca2eb15 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316071, 'name': PowerOnVM_Task, 'duration_secs': 0.355789} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.306440] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e87ffa1-9606-42a3-87ca-4256eca2eb15 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1206.306666] env[61868]: DEBUG nova.compute.manager [None req-8e87ffa1-9606-42a3-87ca-4256eca2eb15 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1206.307426] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5b0ccca-3173-4606-b9ae-07defdfef48e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.664685] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f620f33b-3158-4e43-8ff6-fa5454e6ba60 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "f2d864f1-cd11-4b62-857d-789cf045f22c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1207.664992] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f620f33b-3158-4e43-8ff6-fa5454e6ba60 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "f2d864f1-cd11-4b62-857d-789cf045f22c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1207.665226] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f620f33b-3158-4e43-8ff6-fa5454e6ba60 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "f2d864f1-cd11-4b62-857d-789cf045f22c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1207.665804] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f620f33b-3158-4e43-8ff6-fa5454e6ba60 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "f2d864f1-cd11-4b62-857d-789cf045f22c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1207.665804] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f620f33b-3158-4e43-8ff6-fa5454e6ba60 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "f2d864f1-cd11-4b62-857d-789cf045f22c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1207.670339] env[61868]: INFO nova.compute.manager [None req-f620f33b-3158-4e43-8ff6-fa5454e6ba60 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Terminating instance [ 1207.671936] env[61868]: DEBUG nova.compute.manager [None req-f620f33b-3158-4e43-8ff6-fa5454e6ba60 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1207.671936] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f620f33b-3158-4e43-8ff6-fa5454e6ba60 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1207.671936] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f72b476e-81aa-4cb4-9b4e-0b855a94126f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.680148] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-f620f33b-3158-4e43-8ff6-fa5454e6ba60 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1207.680397] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f864e1d6-c57c-4b99-94b0-418bb2819569 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.687184] env[61868]: DEBUG oslo_vmware.api [None req-f620f33b-3158-4e43-8ff6-fa5454e6ba60 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1207.687184] env[61868]: value = "task-1316072" [ 1207.687184] env[61868]: _type = "Task" [ 1207.687184] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.695191] env[61868]: DEBUG oslo_vmware.api [None req-f620f33b-3158-4e43-8ff6-fa5454e6ba60 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316072, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.196923] env[61868]: DEBUG oslo_vmware.api [None req-f620f33b-3158-4e43-8ff6-fa5454e6ba60 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316072, 'name': PowerOffVM_Task, 'duration_secs': 0.156634} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.197209] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-f620f33b-3158-4e43-8ff6-fa5454e6ba60 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1208.197386] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f620f33b-3158-4e43-8ff6-fa5454e6ba60 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1208.197639] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eaa756e1-9c8e-41f5-9ed7-b264596eb674 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.260671] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f620f33b-3158-4e43-8ff6-fa5454e6ba60 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1208.261009] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f620f33b-3158-4e43-8ff6-fa5454e6ba60 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Deleting contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1208.261212] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-f620f33b-3158-4e43-8ff6-fa5454e6ba60 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Deleting the datastore file [datastore2] f2d864f1-cd11-4b62-857d-789cf045f22c {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1208.261476] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-363a3203-c4ff-4bfa-a3eb-c27b255646ed {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.268463] env[61868]: DEBUG oslo_vmware.api [None req-f620f33b-3158-4e43-8ff6-fa5454e6ba60 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1208.268463] env[61868]: value = "task-1316074" [ 1208.268463] env[61868]: _type = "Task" [ 1208.268463] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.276239] env[61868]: DEBUG oslo_vmware.api [None req-f620f33b-3158-4e43-8ff6-fa5454e6ba60 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316074, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.779825] env[61868]: DEBUG oslo_vmware.api [None req-f620f33b-3158-4e43-8ff6-fa5454e6ba60 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316074, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.153976} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.780166] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-f620f33b-3158-4e43-8ff6-fa5454e6ba60 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1208.780375] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f620f33b-3158-4e43-8ff6-fa5454e6ba60 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Deleted contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1208.780560] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f620f33b-3158-4e43-8ff6-fa5454e6ba60 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1208.780738] env[61868]: INFO nova.compute.manager [None req-f620f33b-3158-4e43-8ff6-fa5454e6ba60 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1208.780981] env[61868]: DEBUG oslo.service.loopingcall [None req-f620f33b-3158-4e43-8ff6-fa5454e6ba60 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1208.781194] env[61868]: DEBUG nova.compute.manager [-] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1208.781288] env[61868]: DEBUG nova.network.neutron [-] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1208.982991] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a5ff6fd1-213a-46bb-b3e5-5fd99a44e815 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1208.983258] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a5ff6fd1-213a-46bb-b3e5-5fd99a44e815 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1209.241072] env[61868]: DEBUG nova.compute.manager [req-ef76bd01-495e-4b54-8f11-1a8e7c7be01e req-df3d3b72-32e6-49f8-b02c-ab97e119c758 service nova] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Received event network-vif-deleted-4c07a7f4-8d47-494b-977a-1aa62477db98 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1209.241332] env[61868]: INFO nova.compute.manager [req-ef76bd01-495e-4b54-8f11-1a8e7c7be01e req-df3d3b72-32e6-49f8-b02c-ab97e119c758 service nova] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Neutron deleted interface 4c07a7f4-8d47-494b-977a-1aa62477db98; detaching it from the instance and deleting it from the info cache [ 1209.241500] env[61868]: DEBUG nova.network.neutron [req-ef76bd01-495e-4b54-8f11-1a8e7c7be01e req-df3d3b72-32e6-49f8-b02c-ab97e119c758 service nova] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1209.486806] env[61868]: DEBUG nova.compute.utils [None req-a5ff6fd1-213a-46bb-b3e5-5fd99a44e815 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1209.720878] env[61868]: DEBUG nova.network.neutron [-] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1209.743817] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a7449446-9772-49f0-88bb-6605205e45d5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.753939] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e31405b-2f33-4854-9a72-abdea251e61e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.776723] env[61868]: DEBUG nova.compute.manager [req-ef76bd01-495e-4b54-8f11-1a8e7c7be01e req-df3d3b72-32e6-49f8-b02c-ab97e119c758 service nova] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Detach interface failed, port_id=4c07a7f4-8d47-494b-977a-1aa62477db98, reason: Instance f2d864f1-cd11-4b62-857d-789cf045f22c could not be found. {{(pid=61868) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1209.991262] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a5ff6fd1-213a-46bb-b3e5-5fd99a44e815 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1210.223731] env[61868]: INFO nova.compute.manager [-] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Took 1.44 seconds to deallocate network for instance. [ 1210.730452] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f620f33b-3158-4e43-8ff6-fa5454e6ba60 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1210.730744] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f620f33b-3158-4e43-8ff6-fa5454e6ba60 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1210.730941] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f620f33b-3158-4e43-8ff6-fa5454e6ba60 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1210.755273] env[61868]: INFO nova.scheduler.client.report [None req-f620f33b-3158-4e43-8ff6-fa5454e6ba60 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Deleted allocations for instance f2d864f1-cd11-4b62-857d-789cf045f22c [ 1211.052062] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a5ff6fd1-213a-46bb-b3e5-5fd99a44e815 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1211.052415] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a5ff6fd1-213a-46bb-b3e5-5fd99a44e815 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1211.054190] env[61868]: INFO nova.compute.manager [None req-a5ff6fd1-213a-46bb-b3e5-5fd99a44e815 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Attaching volume 622da5ad-c85c-484a-8297-2a235bd891eb to /dev/sdb [ 1211.084724] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4629e09c-3317-47e5-92f8-c6264e05c7e1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.091961] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff79365d-056c-4fa0-99a0-2a3b177b74ff {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.104652] env[61868]: DEBUG nova.virt.block_device [None req-a5ff6fd1-213a-46bb-b3e5-5fd99a44e815 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Updating existing volume attachment record: 677b773f-b048-4cea-97ab-d365d181052d {{(pid=61868) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1211.263767] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f620f33b-3158-4e43-8ff6-fa5454e6ba60 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "f2d864f1-cd11-4b62-857d-789cf045f22c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.599s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1212.749733] env[61868]: DEBUG oslo_concurrency.lockutils [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "e18dfd04-5a37-44d1-930b-52b50362755b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1212.750064] env[61868]: DEBUG oslo_concurrency.lockutils [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "e18dfd04-5a37-44d1-930b-52b50362755b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1213.074931] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1213.252016] env[61868]: DEBUG nova.compute.manager [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1213.771239] env[61868]: DEBUG oslo_concurrency.lockutils [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1213.771498] env[61868]: DEBUG oslo_concurrency.lockutils [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1213.772981] env[61868]: INFO nova.compute.claims [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1214.074823] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1214.074991] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61868) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1214.823865] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5b01abc-8a9c-4e40-8a42-ed951eed2a34 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.831712] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50948749-22c7-4f53-9a4c-b309ffeb22ff {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.862031] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-607e53c7-e354-414d-8644-051461b3e8ea {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.869117] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a49f0928-3518-4f11-90c6-564dd4cd60e1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.881843] env[61868]: DEBUG nova.compute.provider_tree [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1215.385468] env[61868]: DEBUG nova.scheduler.client.report [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1215.890467] env[61868]: DEBUG oslo_concurrency.lockutils [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.119s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1215.891037] env[61868]: DEBUG nova.compute.manager [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1216.075033] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager.update_available_resource {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1216.395848] env[61868]: DEBUG nova.compute.utils [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1216.397255] env[61868]: DEBUG nova.compute.manager [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1216.397423] env[61868]: DEBUG nova.network.neutron [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1216.452823] env[61868]: DEBUG nova.policy [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4a981d6292f34cd7aba2108a9b9f3a90', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '28a7014141b24101ba7920bc83be4a9e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 1216.577899] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1216.578181] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1216.578364] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1216.578529] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61868) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1216.579483] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a6a8c7f-84d0-4b7a-8731-27b91666c823 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.587917] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8427ede-c438-4399-ad75-09734b5d0762 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.601560] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e8f0aa3-9543-4cce-9521-ffa92d76384f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.607908] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ed8b112-e9a4-47b3-ba3d-6808b205fa0a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.636933] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180865MB free_disk=134GB free_vcpus=48 pci_devices=None {{(pid=61868) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1216.637085] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1216.637272] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1216.775378] env[61868]: DEBUG nova.network.neutron [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Successfully created port: 644d55b3-8c23-4a45-93d6-9d5464d1631e {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1216.900545] env[61868]: DEBUG nova.compute.manager [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1217.662646] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5ff6fd1-213a-46bb-b3e5-5fd99a44e815 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Volume attach. Driver type: vmdk {{(pid=61868) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1217.662736] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5ff6fd1-213a-46bb-b3e5-5fd99a44e815 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281640', 'volume_id': '622da5ad-c85c-484a-8297-2a235bd891eb', 'name': 'volume-622da5ad-c85c-484a-8297-2a235bd891eb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e', 'attached_at': '', 'detached_at': '', 'volume_id': '622da5ad-c85c-484a-8297-2a235bd891eb', 'serial': '622da5ad-c85c-484a-8297-2a235bd891eb'} {{(pid=61868) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1217.663758] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e43425eb-12b5-4b86-a8e3-f4e8e2a9f100 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.667354] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 59171bf1-f2d4-4ff5-a661-025f5dca7600 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1217.667495] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1217.667681] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance e18dfd04-5a37-44d1-930b-52b50362755b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1217.667791] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=61868) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1217.667925] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=61868) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1217.685602] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eecb029-6474-4f21-b404-93bb7a6c68da {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.709790] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5ff6fd1-213a-46bb-b3e5-5fd99a44e815 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] volume-622da5ad-c85c-484a-8297-2a235bd891eb/volume-622da5ad-c85c-484a-8297-2a235bd891eb.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1217.712128] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2dd11136-cc6f-4c2e-b54b-16f9cd59390d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.731362] env[61868]: DEBUG oslo_vmware.api [None req-a5ff6fd1-213a-46bb-b3e5-5fd99a44e815 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1217.731362] env[61868]: value = "task-1316079" [ 1217.731362] env[61868]: _type = "Task" [ 1217.731362] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1217.742956] env[61868]: DEBUG oslo_vmware.api [None req-a5ff6fd1-213a-46bb-b3e5-5fd99a44e815 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316079, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.747431] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71ed3725-0444-4cf1-817b-e01b6b990e2b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.753806] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cde9cbd-4fa8-4879-b07e-b70d8618865e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.783332] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b8ec9af-20c5-4025-a8f9-d0385104ba43 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.790367] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11650220-e938-44e9-8603-2321cf59702e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.803365] env[61868]: DEBUG nova.compute.provider_tree [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1217.911120] env[61868]: DEBUG nova.compute.manager [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1217.936304] env[61868]: DEBUG nova.virt.hardware [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1217.936576] env[61868]: DEBUG nova.virt.hardware [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1217.936817] env[61868]: DEBUG nova.virt.hardware [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1217.937045] env[61868]: DEBUG nova.virt.hardware [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1217.937205] env[61868]: DEBUG nova.virt.hardware [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1217.937357] env[61868]: DEBUG nova.virt.hardware [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1217.937634] env[61868]: DEBUG nova.virt.hardware [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1217.937825] env[61868]: DEBUG nova.virt.hardware [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1217.938039] env[61868]: DEBUG nova.virt.hardware [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1217.938290] env[61868]: DEBUG nova.virt.hardware [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1217.938483] env[61868]: DEBUG nova.virt.hardware [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1217.939409] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dadbd54-48fd-48d6-86d0-c6b20796c731 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.947333] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecb9d879-0d1d-447f-b84f-f3f9f640b825 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.147842] env[61868]: DEBUG nova.compute.manager [req-41a36778-0302-4ac7-b798-fd5bc9f08e10 req-c61b6a6b-57f7-4c85-bac9-d20dd6d71c2b service nova] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Received event network-vif-plugged-644d55b3-8c23-4a45-93d6-9d5464d1631e {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1218.148047] env[61868]: DEBUG oslo_concurrency.lockutils [req-41a36778-0302-4ac7-b798-fd5bc9f08e10 req-c61b6a6b-57f7-4c85-bac9-d20dd6d71c2b service nova] Acquiring lock "e18dfd04-5a37-44d1-930b-52b50362755b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1218.148269] env[61868]: DEBUG oslo_concurrency.lockutils [req-41a36778-0302-4ac7-b798-fd5bc9f08e10 req-c61b6a6b-57f7-4c85-bac9-d20dd6d71c2b service nova] Lock "e18dfd04-5a37-44d1-930b-52b50362755b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1218.148441] env[61868]: DEBUG oslo_concurrency.lockutils [req-41a36778-0302-4ac7-b798-fd5bc9f08e10 req-c61b6a6b-57f7-4c85-bac9-d20dd6d71c2b service nova] Lock "e18dfd04-5a37-44d1-930b-52b50362755b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1218.148615] env[61868]: DEBUG nova.compute.manager [req-41a36778-0302-4ac7-b798-fd5bc9f08e10 req-c61b6a6b-57f7-4c85-bac9-d20dd6d71c2b service nova] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] No waiting events found dispatching network-vif-plugged-644d55b3-8c23-4a45-93d6-9d5464d1631e {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1218.148782] env[61868]: WARNING nova.compute.manager [req-41a36778-0302-4ac7-b798-fd5bc9f08e10 req-c61b6a6b-57f7-4c85-bac9-d20dd6d71c2b service nova] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Received unexpected event network-vif-plugged-644d55b3-8c23-4a45-93d6-9d5464d1631e for instance with vm_state building and task_state spawning. [ 1218.226126] env[61868]: DEBUG nova.network.neutron [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Successfully updated port: 644d55b3-8c23-4a45-93d6-9d5464d1631e {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1218.241950] env[61868]: DEBUG oslo_vmware.api [None req-a5ff6fd1-213a-46bb-b3e5-5fd99a44e815 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316079, 'name': ReconfigVM_Task, 'duration_secs': 0.327741} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1218.242278] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5ff6fd1-213a-46bb-b3e5-5fd99a44e815 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Reconfigured VM instance instance-0000006b to attach disk [datastore1] volume-622da5ad-c85c-484a-8297-2a235bd891eb/volume-622da5ad-c85c-484a-8297-2a235bd891eb.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1218.247067] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cb6ab72a-0a2b-448f-8771-210ddf70afdd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.261466] env[61868]: DEBUG oslo_vmware.api [None req-a5ff6fd1-213a-46bb-b3e5-5fd99a44e815 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1218.261466] env[61868]: value = "task-1316080" [ 1218.261466] env[61868]: _type = "Task" [ 1218.261466] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1218.270381] env[61868]: DEBUG oslo_vmware.api [None req-a5ff6fd1-213a-46bb-b3e5-5fd99a44e815 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316080, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.306403] env[61868]: DEBUG nova.scheduler.client.report [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1218.728708] env[61868]: DEBUG oslo_concurrency.lockutils [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "refresh_cache-e18dfd04-5a37-44d1-930b-52b50362755b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1218.728876] env[61868]: DEBUG oslo_concurrency.lockutils [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquired lock "refresh_cache-e18dfd04-5a37-44d1-930b-52b50362755b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1218.729039] env[61868]: DEBUG nova.network.neutron [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1218.771971] env[61868]: DEBUG oslo_vmware.api [None req-a5ff6fd1-213a-46bb-b3e5-5fd99a44e815 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316080, 'name': ReconfigVM_Task, 'duration_secs': 0.128378} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1218.772284] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5ff6fd1-213a-46bb-b3e5-5fd99a44e815 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281640', 'volume_id': '622da5ad-c85c-484a-8297-2a235bd891eb', 'name': 'volume-622da5ad-c85c-484a-8297-2a235bd891eb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e', 'attached_at': '', 'detached_at': '', 'volume_id': '622da5ad-c85c-484a-8297-2a235bd891eb', 'serial': '622da5ad-c85c-484a-8297-2a235bd891eb'} {{(pid=61868) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1218.810824] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61868) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1218.810991] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.174s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1219.260592] env[61868]: DEBUG nova.network.neutron [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1219.383742] env[61868]: DEBUG nova.network.neutron [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Updating instance_info_cache with network_info: [{"id": "644d55b3-8c23-4a45-93d6-9d5464d1631e", "address": "fa:16:3e:db:75:9d", "network": {"id": "7e8cbecf-dfdd-43f2-8745-ae1d2d55d309", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-33935587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a7014141b24101ba7920bc83be4a9e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4db2961d-273d-4634-9d06-a94fa9d384fb", "external-id": "nsx-vlan-transportzone-572", "segmentation_id": 572, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap644d55b3-8c", "ovs_interfaceid": "644d55b3-8c23-4a45-93d6-9d5464d1631e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1219.807645] env[61868]: DEBUG nova.objects.instance [None req-a5ff6fd1-213a-46bb-b3e5-5fd99a44e815 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lazy-loading 'flavor' on Instance uuid 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1219.811274] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1219.811274] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1219.811274] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Starting heal instance info cache {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1219.887046] env[61868]: DEBUG oslo_concurrency.lockutils [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Releasing lock "refresh_cache-e18dfd04-5a37-44d1-930b-52b50362755b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1219.887046] env[61868]: DEBUG nova.compute.manager [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Instance network_info: |[{"id": "644d55b3-8c23-4a45-93d6-9d5464d1631e", "address": "fa:16:3e:db:75:9d", "network": {"id": "7e8cbecf-dfdd-43f2-8745-ae1d2d55d309", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-33935587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a7014141b24101ba7920bc83be4a9e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4db2961d-273d-4634-9d06-a94fa9d384fb", "external-id": "nsx-vlan-transportzone-572", "segmentation_id": 572, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap644d55b3-8c", "ovs_interfaceid": "644d55b3-8c23-4a45-93d6-9d5464d1631e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1219.887290] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:db:75:9d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4db2961d-273d-4634-9d06-a94fa9d384fb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '644d55b3-8c23-4a45-93d6-9d5464d1631e', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1219.894381] env[61868]: DEBUG oslo.service.loopingcall [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1219.894591] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1219.894810] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a29d1561-6908-4fce-99fe-e392b0b9bc11 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.914186] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1219.914186] env[61868]: value = "task-1316081" [ 1219.914186] env[61868]: _type = "Task" [ 1219.914186] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1219.925474] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1316081, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.176317] env[61868]: DEBUG nova.compute.manager [req-9abb6890-6e9b-4486-985a-2847bc4061f7 req-f3c36c5a-fc1f-4486-87e0-d70d42677ab4 service nova] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Received event network-changed-644d55b3-8c23-4a45-93d6-9d5464d1631e {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1220.176663] env[61868]: DEBUG nova.compute.manager [req-9abb6890-6e9b-4486-985a-2847bc4061f7 req-f3c36c5a-fc1f-4486-87e0-d70d42677ab4 service nova] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Refreshing instance network info cache due to event network-changed-644d55b3-8c23-4a45-93d6-9d5464d1631e. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1220.177167] env[61868]: DEBUG oslo_concurrency.lockutils [req-9abb6890-6e9b-4486-985a-2847bc4061f7 req-f3c36c5a-fc1f-4486-87e0-d70d42677ab4 service nova] Acquiring lock "refresh_cache-e18dfd04-5a37-44d1-930b-52b50362755b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1220.177423] env[61868]: DEBUG oslo_concurrency.lockutils [req-9abb6890-6e9b-4486-985a-2847bc4061f7 req-f3c36c5a-fc1f-4486-87e0-d70d42677ab4 service nova] Acquired lock "refresh_cache-e18dfd04-5a37-44d1-930b-52b50362755b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1220.177780] env[61868]: DEBUG nova.network.neutron [req-9abb6890-6e9b-4486-985a-2847bc4061f7 req-f3c36c5a-fc1f-4486-87e0-d70d42677ab4 service nova] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Refreshing network info cache for port 644d55b3-8c23-4a45-93d6-9d5464d1631e {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1220.315278] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a5ff6fd1-213a-46bb-b3e5-5fd99a44e815 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 9.262s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1220.424885] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1316081, 'name': CreateVM_Task, 'duration_secs': 0.292311} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1220.425052] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1220.425677] env[61868]: DEBUG oslo_concurrency.lockutils [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1220.425860] env[61868]: DEBUG oslo_concurrency.lockutils [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1220.426203] env[61868]: DEBUG oslo_concurrency.lockutils [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1220.426459] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-911a36a6-6aec-4071-89ee-3eb299c16800 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.430490] env[61868]: DEBUG oslo_vmware.api [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1220.430490] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5210c379-3ffa-bff5-72ca-46285415d500" [ 1220.430490] env[61868]: _type = "Task" [ 1220.430490] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1220.437636] env[61868]: DEBUG oslo_vmware.api [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5210c379-3ffa-bff5-72ca-46285415d500, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.720475] env[61868]: INFO nova.compute.manager [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Rebuilding instance [ 1220.763556] env[61868]: DEBUG nova.compute.manager [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1220.764412] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d4876cb-d1fd-41b3-a87b-6fb8e5294acc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.817302] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Didn't find any instances for network info cache update. {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1220.817739] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1220.817913] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1220.818077] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1220.818225] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1220.877701] env[61868]: DEBUG nova.network.neutron [req-9abb6890-6e9b-4486-985a-2847bc4061f7 req-f3c36c5a-fc1f-4486-87e0-d70d42677ab4 service nova] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Updated VIF entry in instance network info cache for port 644d55b3-8c23-4a45-93d6-9d5464d1631e. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1220.878093] env[61868]: DEBUG nova.network.neutron [req-9abb6890-6e9b-4486-985a-2847bc4061f7 req-f3c36c5a-fc1f-4486-87e0-d70d42677ab4 service nova] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Updating instance_info_cache with network_info: [{"id": "644d55b3-8c23-4a45-93d6-9d5464d1631e", "address": "fa:16:3e:db:75:9d", "network": {"id": "7e8cbecf-dfdd-43f2-8745-ae1d2d55d309", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-33935587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a7014141b24101ba7920bc83be4a9e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4db2961d-273d-4634-9d06-a94fa9d384fb", "external-id": "nsx-vlan-transportzone-572", "segmentation_id": 572, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap644d55b3-8c", "ovs_interfaceid": "644d55b3-8c23-4a45-93d6-9d5464d1631e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1220.940833] env[61868]: DEBUG oslo_vmware.api [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5210c379-3ffa-bff5-72ca-46285415d500, 'name': SearchDatastore_Task, 'duration_secs': 0.009198} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1220.941170] env[61868]: DEBUG oslo_concurrency.lockutils [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1220.941418] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1220.941657] env[61868]: DEBUG oslo_concurrency.lockutils [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1220.941809] env[61868]: DEBUG oslo_concurrency.lockutils [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1220.941989] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1220.942263] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2b5d83e1-6e44-4a44-b6f6-6d87218d0c4c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.950482] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1220.950647] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1220.951585] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7ce5de33-cb3d-4e68-a866-3bc2b0e904e0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.956480] env[61868]: DEBUG oslo_vmware.api [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1220.956480] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52bf3df3-fa20-595a-600d-ef0b123764a4" [ 1220.956480] env[61868]: _type = "Task" [ 1220.956480] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1220.963650] env[61868]: DEBUG oslo_vmware.api [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52bf3df3-fa20-595a-600d-ef0b123764a4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.277214] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1221.277475] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bf8233d6-728b-4b38-9761-28668017f276 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.284439] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1221.284439] env[61868]: value = "task-1316082" [ 1221.284439] env[61868]: _type = "Task" [ 1221.284439] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.292103] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316082, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.381155] env[61868]: DEBUG oslo_concurrency.lockutils [req-9abb6890-6e9b-4486-985a-2847bc4061f7 req-f3c36c5a-fc1f-4486-87e0-d70d42677ab4 service nova] Releasing lock "refresh_cache-e18dfd04-5a37-44d1-930b-52b50362755b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1221.465981] env[61868]: DEBUG oslo_vmware.api [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52bf3df3-fa20-595a-600d-ef0b123764a4, 'name': SearchDatastore_Task, 'duration_secs': 0.007601} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1221.466772] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d911cde3-32ca-4ea8-bd7a-cf05ebc5537e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.471763] env[61868]: DEBUG oslo_vmware.api [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1221.471763] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52d1a1e5-a4e8-9506-98e5-84719b21d464" [ 1221.471763] env[61868]: _type = "Task" [ 1221.471763] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.478745] env[61868]: DEBUG oslo_vmware.api [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52d1a1e5-a4e8-9506-98e5-84719b21d464, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.793946] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316082, 'name': PowerOffVM_Task, 'duration_secs': 0.173848} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1221.794381] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1221.845311] env[61868]: INFO nova.compute.manager [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Detaching volume 622da5ad-c85c-484a-8297-2a235bd891eb [ 1221.874531] env[61868]: INFO nova.virt.block_device [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Attempting to driver detach volume 622da5ad-c85c-484a-8297-2a235bd891eb from mountpoint /dev/sdb [ 1221.874856] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Volume detach. Driver type: vmdk {{(pid=61868) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1221.875105] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281640', 'volume_id': '622da5ad-c85c-484a-8297-2a235bd891eb', 'name': 'volume-622da5ad-c85c-484a-8297-2a235bd891eb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e', 'attached_at': '', 'detached_at': '', 'volume_id': '622da5ad-c85c-484a-8297-2a235bd891eb', 'serial': '622da5ad-c85c-484a-8297-2a235bd891eb'} {{(pid=61868) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1221.876095] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94e1fd29-998c-4bac-b017-fa12e327349e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.897852] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b121283-c582-4dca-976f-afd7ec64a30f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.904991] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2be5a01c-eef4-418f-a0c5-7c71c6a376af {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.926305] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37af451b-ed2b-498e-aed4-1eec0bf13558 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.941662] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] The volume has not been displaced from its original location: [datastore1] volume-622da5ad-c85c-484a-8297-2a235bd891eb/volume-622da5ad-c85c-484a-8297-2a235bd891eb.vmdk. No consolidation needed. {{(pid=61868) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1221.946898] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Reconfiguring VM instance instance-0000006b to detach disk 2001 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1221.947227] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1bb39ed7-76f0-4118-b9ba-147a21ef9b58 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.965571] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1221.965571] env[61868]: value = "task-1316083" [ 1221.965571] env[61868]: _type = "Task" [ 1221.965571] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.973435] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316083, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.980956] env[61868]: DEBUG oslo_vmware.api [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52d1a1e5-a4e8-9506-98e5-84719b21d464, 'name': SearchDatastore_Task, 'duration_secs': 0.008654} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1221.981199] env[61868]: DEBUG oslo_concurrency.lockutils [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1221.981451] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] e18dfd04-5a37-44d1-930b-52b50362755b/e18dfd04-5a37-44d1-930b-52b50362755b.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1221.981687] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-12b41d24-d055-4e27-a270-12506d943a07 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.986990] env[61868]: DEBUG oslo_vmware.api [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1221.986990] env[61868]: value = "task-1316084" [ 1221.986990] env[61868]: _type = "Task" [ 1221.986990] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.996941] env[61868]: DEBUG oslo_vmware.api [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316084, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.474998] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316083, 'name': ReconfigVM_Task, 'duration_secs': 0.178884} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1222.475328] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Reconfigured VM instance instance-0000006b to detach disk 2001 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1222.479961] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-141d2a1a-fe6b-4e1f-aa6b-ebb840a7af14 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.497705] env[61868]: DEBUG oslo_vmware.api [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316084, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.498972] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1222.498972] env[61868]: value = "task-1316085" [ 1222.498972] env[61868]: _type = "Task" [ 1222.498972] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1222.506295] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316085, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.997868] env[61868]: DEBUG oslo_vmware.api [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316084, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.007275] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316085, 'name': ReconfigVM_Task, 'duration_secs': 0.478494} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1223.007569] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281640', 'volume_id': '622da5ad-c85c-484a-8297-2a235bd891eb', 'name': 'volume-622da5ad-c85c-484a-8297-2a235bd891eb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e', 'attached_at': '', 'detached_at': '', 'volume_id': '622da5ad-c85c-484a-8297-2a235bd891eb', 'serial': '622da5ad-c85c-484a-8297-2a235bd891eb'} {{(pid=61868) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1223.499454] env[61868]: DEBUG oslo_vmware.api [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316084, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.999528] env[61868]: DEBUG oslo_vmware.api [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316084, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.054056] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1224.054056] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e0ed108c-44eb-43fd-9f11-e12afea19efb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.060853] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1224.060853] env[61868]: value = "task-1316086" [ 1224.060853] env[61868]: _type = "Task" [ 1224.060853] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.070220] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] VM already powered off {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1224.070422] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Volume detach. Driver type: vmdk {{(pid=61868) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1224.070612] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281640', 'volume_id': '622da5ad-c85c-484a-8297-2a235bd891eb', 'name': 'volume-622da5ad-c85c-484a-8297-2a235bd891eb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e', 'attached_at': '', 'detached_at': '', 'volume_id': '622da5ad-c85c-484a-8297-2a235bd891eb', 'serial': '622da5ad-c85c-484a-8297-2a235bd891eb'} {{(pid=61868) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1224.071306] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c214907b-78b2-4e72-8a64-ed5241ed19ea {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.088047] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1224.090181] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e92bebd1-961d-4c0c-ba38-0ce277d7d560 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.096447] env[61868]: WARNING nova.virt.vmwareapi.driver [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] The volume None does not exist!: nova.exception.DiskNotFound: Unable to find volume [ 1224.096624] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1224.097340] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7bea5b1-e0ad-40cd-8c70-73069119147b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.104200] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1224.104416] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6928adf4-6a2c-4bef-bc01-00ce159234a9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.500455] env[61868]: DEBUG oslo_vmware.api [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316084, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.000945] env[61868]: DEBUG oslo_vmware.api [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316084, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.502121] env[61868]: DEBUG oslo_vmware.api [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316084, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.003253] env[61868]: DEBUG oslo_vmware.api [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316084, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.502960] env[61868]: DEBUG oslo_vmware.api [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316084, 'name': CopyVirtualDisk_Task, 'duration_secs': 4.040516} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1226.503359] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] e18dfd04-5a37-44d1-930b-52b50362755b/e18dfd04-5a37-44d1-930b-52b50362755b.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1226.503450] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1226.503739] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5cdd50cc-38a0-4f29-a6bb-b192add22198 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.510127] env[61868]: DEBUG oslo_vmware.api [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1226.510127] env[61868]: value = "task-1316088" [ 1226.510127] env[61868]: _type = "Task" [ 1226.510127] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.517210] env[61868]: DEBUG oslo_vmware.api [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316088, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.562857] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1226.563131] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1226.563322] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Deleting the datastore file [datastore1] 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1226.563599] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-956f9f87-482f-4030-a2fd-b8de49581388 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.569933] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1226.569933] env[61868]: value = "task-1316089" [ 1226.569933] env[61868]: _type = "Task" [ 1226.569933] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.577973] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316089, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.019582] env[61868]: DEBUG oslo_vmware.api [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316088, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067578} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1227.019873] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1227.020668] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-667dc819-65df-433d-850b-2da8d3accc78 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.043382] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] e18dfd04-5a37-44d1-930b-52b50362755b/e18dfd04-5a37-44d1-930b-52b50362755b.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1227.044559] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9a24c667-58a7-434f-87b2-d4249bb06fe6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.063254] env[61868]: DEBUG oslo_vmware.api [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1227.063254] env[61868]: value = "task-1316090" [ 1227.063254] env[61868]: _type = "Task" [ 1227.063254] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1227.071900] env[61868]: DEBUG oslo_vmware.api [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316090, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.080822] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316089, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.573415] env[61868]: DEBUG oslo_vmware.api [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316090, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.581193] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316089, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.526787} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1227.581514] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1227.581752] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1227.581967] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1228.074361] env[61868]: DEBUG oslo_vmware.api [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316090, 'name': ReconfigVM_Task, 'duration_secs': 0.612258} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.074640] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Reconfigured VM instance instance-0000006c to attach disk [datastore1] e18dfd04-5a37-44d1-930b-52b50362755b/e18dfd04-5a37-44d1-930b-52b50362755b.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1228.075326] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-33af9297-51fa-480b-84ff-98bee8b4d617 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.081818] env[61868]: DEBUG oslo_vmware.api [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1228.081818] env[61868]: value = "task-1316091" [ 1228.081818] env[61868]: _type = "Task" [ 1228.081818] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1228.086555] env[61868]: INFO nova.virt.block_device [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Booting with volume 622da5ad-c85c-484a-8297-2a235bd891eb at /dev/sdb [ 1228.091463] env[61868]: DEBUG oslo_vmware.api [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316091, 'name': Rename_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.121242] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e96dd25f-3c65-409d-94b3-c99891be4fbe {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.129725] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43ed9f08-b140-44fc-ab7f-ff90fed2a907 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.153308] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c038275d-586a-4e3d-ab78-8556112fa57d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.160945] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fa94a69-ae76-4a6b-98cd-cf1aae734c83 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.185961] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b16f4da-1c77-4eda-8625-92cd8546e22f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.195075] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f08fd68e-14a1-4920-a77f-d2c0fa7d42b6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.208775] env[61868]: DEBUG nova.virt.block_device [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Updating existing volume attachment record: 2a893768-6df9-48a3-b047-e4206c9b659d {{(pid=61868) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1228.592994] env[61868]: DEBUG oslo_vmware.api [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316091, 'name': Rename_Task, 'duration_secs': 0.136713} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.593302] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1228.593538] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bbdf9b15-ba24-4956-be6f-f7d70dcc41a4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.599344] env[61868]: DEBUG oslo_vmware.api [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1228.599344] env[61868]: value = "task-1316092" [ 1228.599344] env[61868]: _type = "Task" [ 1228.599344] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1228.606209] env[61868]: DEBUG oslo_vmware.api [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316092, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.108899] env[61868]: DEBUG oslo_vmware.api [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316092, 'name': PowerOnVM_Task, 'duration_secs': 0.437712} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1229.109132] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1229.109339] env[61868]: INFO nova.compute.manager [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Took 11.20 seconds to spawn the instance on the hypervisor. [ 1229.109520] env[61868]: DEBUG nova.compute.manager [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1229.110264] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ec2f353-954a-447c-821f-f252a0225da2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.627875] env[61868]: INFO nova.compute.manager [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Took 15.87 seconds to build instance. [ 1230.129832] env[61868]: DEBUG oslo_concurrency.lockutils [None req-89f257b2-0b81-4a6e-96c0-aa7da2deac9d tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "e18dfd04-5a37-44d1-930b-52b50362755b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.380s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1230.155940] env[61868]: DEBUG nova.compute.manager [req-44ff440d-7284-45cc-b35d-9f8d9f9714d6 req-0d38b947-69db-4f57-a582-9367c987fc4b service nova] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Received event network-changed-644d55b3-8c23-4a45-93d6-9d5464d1631e {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1230.156169] env[61868]: DEBUG nova.compute.manager [req-44ff440d-7284-45cc-b35d-9f8d9f9714d6 req-0d38b947-69db-4f57-a582-9367c987fc4b service nova] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Refreshing instance network info cache due to event network-changed-644d55b3-8c23-4a45-93d6-9d5464d1631e. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1230.156400] env[61868]: DEBUG oslo_concurrency.lockutils [req-44ff440d-7284-45cc-b35d-9f8d9f9714d6 req-0d38b947-69db-4f57-a582-9367c987fc4b service nova] Acquiring lock "refresh_cache-e18dfd04-5a37-44d1-930b-52b50362755b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1230.156548] env[61868]: DEBUG oslo_concurrency.lockutils [req-44ff440d-7284-45cc-b35d-9f8d9f9714d6 req-0d38b947-69db-4f57-a582-9367c987fc4b service nova] Acquired lock "refresh_cache-e18dfd04-5a37-44d1-930b-52b50362755b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1230.156727] env[61868]: DEBUG nova.network.neutron [req-44ff440d-7284-45cc-b35d-9f8d9f9714d6 req-0d38b947-69db-4f57-a582-9367c987fc4b service nova] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Refreshing network info cache for port 644d55b3-8c23-4a45-93d6-9d5464d1631e {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1230.323285] env[61868]: DEBUG nova.virt.hardware [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1230.323720] env[61868]: DEBUG nova.virt.hardware [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1230.323799] env[61868]: DEBUG nova.virt.hardware [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1230.323934] env[61868]: DEBUG nova.virt.hardware [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1230.324112] env[61868]: DEBUG nova.virt.hardware [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1230.324267] env[61868]: DEBUG nova.virt.hardware [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1230.324488] env[61868]: DEBUG nova.virt.hardware [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1230.324651] env[61868]: DEBUG nova.virt.hardware [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1230.324833] env[61868]: DEBUG nova.virt.hardware [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1230.324999] env[61868]: DEBUG nova.virt.hardware [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1230.325191] env[61868]: DEBUG nova.virt.hardware [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1230.326089] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66eb8edf-374c-4f40-9219-14e2f0e670d8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.336553] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91eeba69-e215-4e7a-94ff-1bf25f6c9e3e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.350032] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d7:73:72', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd298db54-f13d-4bf6-b6c2-755074b3047f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2bac9dd8-51e8-4aad-bf36-90e045aa9106', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1230.357318] env[61868]: DEBUG oslo.service.loopingcall [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1230.357577] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1230.357789] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fd2315ba-8c44-4cb4-b669-8164f7163226 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.376486] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1230.376486] env[61868]: value = "task-1316093" [ 1230.376486] env[61868]: _type = "Task" [ 1230.376486] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1230.383721] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1316093, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.886755] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1316093, 'name': CreateVM_Task, 'duration_secs': 0.305224} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1230.887100] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1230.887628] env[61868]: DEBUG oslo_concurrency.lockutils [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1230.887800] env[61868]: DEBUG oslo_concurrency.lockutils [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1230.888139] env[61868]: DEBUG oslo_concurrency.lockutils [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1230.888389] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-15e939bf-c873-4719-abfb-0e8112ae8c53 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.891267] env[61868]: DEBUG nova.network.neutron [req-44ff440d-7284-45cc-b35d-9f8d9f9714d6 req-0d38b947-69db-4f57-a582-9367c987fc4b service nova] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Updated VIF entry in instance network info cache for port 644d55b3-8c23-4a45-93d6-9d5464d1631e. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1230.891620] env[61868]: DEBUG nova.network.neutron [req-44ff440d-7284-45cc-b35d-9f8d9f9714d6 req-0d38b947-69db-4f57-a582-9367c987fc4b service nova] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Updating instance_info_cache with network_info: [{"id": "644d55b3-8c23-4a45-93d6-9d5464d1631e", "address": "fa:16:3e:db:75:9d", "network": {"id": "7e8cbecf-dfdd-43f2-8745-ae1d2d55d309", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-33935587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a7014141b24101ba7920bc83be4a9e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4db2961d-273d-4634-9d06-a94fa9d384fb", "external-id": "nsx-vlan-transportzone-572", "segmentation_id": 572, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap644d55b3-8c", "ovs_interfaceid": "644d55b3-8c23-4a45-93d6-9d5464d1631e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1230.893882] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1230.893882] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]520091ac-dcab-7723-d79c-5ebd635276c9" [ 1230.893882] env[61868]: _type = "Task" [ 1230.893882] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1230.901483] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]520091ac-dcab-7723-d79c-5ebd635276c9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.395946] env[61868]: DEBUG oslo_concurrency.lockutils [req-44ff440d-7284-45cc-b35d-9f8d9f9714d6 req-0d38b947-69db-4f57-a582-9367c987fc4b service nova] Releasing lock "refresh_cache-e18dfd04-5a37-44d1-930b-52b50362755b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1231.405636] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]520091ac-dcab-7723-d79c-5ebd635276c9, 'name': SearchDatastore_Task, 'duration_secs': 0.009511} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1231.405985] env[61868]: DEBUG oslo_concurrency.lockutils [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1231.406243] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1231.406478] env[61868]: DEBUG oslo_concurrency.lockutils [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1231.406629] env[61868]: DEBUG oslo_concurrency.lockutils [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1231.406817] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1231.407078] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f7ffa6cd-5e95-45fb-8a38-a0f8024a12af {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.417203] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1231.417374] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1231.418371] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-51661287-d7f7-442b-b72a-2708c2db409f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.423220] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1231.423220] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52209ff4-21d3-3d6e-7276-de7a4389c9f8" [ 1231.423220] env[61868]: _type = "Task" [ 1231.423220] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1231.430639] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52209ff4-21d3-3d6e-7276-de7a4389c9f8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.934212] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52209ff4-21d3-3d6e-7276-de7a4389c9f8, 'name': SearchDatastore_Task, 'duration_secs': 0.007779} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1231.935078] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb2c3e6a-e4b2-4acb-a5b7-5ff97bc11763 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.940775] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1231.940775] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52df77f3-494b-587f-0e54-8b0cfc87f3b1" [ 1231.940775] env[61868]: _type = "Task" [ 1231.940775] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1231.948026] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52df77f3-494b-587f-0e54-8b0cfc87f3b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.450550] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52df77f3-494b-587f-0e54-8b0cfc87f3b1, 'name': SearchDatastore_Task, 'duration_secs': 0.009328} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1232.450838] env[61868]: DEBUG oslo_concurrency.lockutils [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1232.451093] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e/7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1232.451353] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c37c9b4c-5cfd-4969-823e-d01e6b58ba6c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.457062] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1232.457062] env[61868]: value = "task-1316094" [ 1232.457062] env[61868]: _type = "Task" [ 1232.457062] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1232.463843] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316094, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.966636] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316094, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.470918} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1232.967032] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e/7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1232.967231] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1232.967495] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a9454352-bbaa-446e-a379-ebac76169535 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.973081] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1232.973081] env[61868]: value = "task-1316095" [ 1232.973081] env[61868]: _type = "Task" [ 1232.973081] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1232.979934] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316095, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.482495] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316095, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075933} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1233.482708] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1233.483485] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-301dc834-6549-4bfe-918e-9b0cced8a9a9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.504541] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e/7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1233.504789] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0b84a6fb-3a16-4ae2-a723-2f20f63bd695 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.522929] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1233.522929] env[61868]: value = "task-1316096" [ 1233.522929] env[61868]: _type = "Task" [ 1233.522929] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1233.533226] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316096, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.033753] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316096, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.533799] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316096, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.033493] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316096, 'name': ReconfigVM_Task, 'duration_secs': 1.065079} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.033799] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Reconfigured VM instance instance-0000006b to attach disk [datastore1] 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e/7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1235.035176] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'boot_index': 0, 'device_type': 'disk', 'encryption_options': None, 'device_name': '/dev/sda', 'size': 0, 'disk_bus': None, 'encryption_secret_uuid': None, 'encryption_format': None, 'guest_format': None, 'encrypted': False, 'image_id': 'c4fdfa42-c9a0-41ca-801e-4e55c3d4328d'}], 'ephemerals': [], 'block_device_mapping': [{'boot_index': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281640', 'volume_id': '622da5ad-c85c-484a-8297-2a235bd891eb', 'name': 'volume-622da5ad-c85c-484a-8297-2a235bd891eb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e', 'attached_at': '', 'detached_at': '', 'volume_id': '622da5ad-c85c-484a-8297-2a235bd891eb', 'serial': '622da5ad-c85c-484a-8297-2a235bd891eb'}, 'device_type': None, 'delete_on_termination': False, 'disk_bus': None, 'mount_device': '/dev/sdb', 'guest_format': None, 'attachment_id': '2a893768-6df9-48a3-b047-e4206c9b659d', 'volume_type': None}], 'swap': None} {{(pid=61868) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1235.035387] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Volume attach. Driver type: vmdk {{(pid=61868) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1235.035605] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281640', 'volume_id': '622da5ad-c85c-484a-8297-2a235bd891eb', 'name': 'volume-622da5ad-c85c-484a-8297-2a235bd891eb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e', 'attached_at': '', 'detached_at': '', 'volume_id': '622da5ad-c85c-484a-8297-2a235bd891eb', 'serial': '622da5ad-c85c-484a-8297-2a235bd891eb'} {{(pid=61868) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1235.036340] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-413af08f-2161-4be2-b48d-d48c2e775865 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.050943] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dc62748-394a-4907-8e64-5d0385ac6cdf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.073441] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] volume-622da5ad-c85c-484a-8297-2a235bd891eb/volume-622da5ad-c85c-484a-8297-2a235bd891eb.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1235.073660] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5d774e83-715b-41c5-b048-856f3a6856e3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.089948] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1235.089948] env[61868]: value = "task-1316097" [ 1235.089948] env[61868]: _type = "Task" [ 1235.089948] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.096968] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316097, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.599339] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316097, 'name': ReconfigVM_Task, 'duration_secs': 0.269921} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.599611] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Reconfigured VM instance instance-0000006b to attach disk [datastore1] volume-622da5ad-c85c-484a-8297-2a235bd891eb/volume-622da5ad-c85c-484a-8297-2a235bd891eb.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1235.604223] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1cacc58f-91d4-4083-b1af-cbf4ee7f7ecb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.617792] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1235.617792] env[61868]: value = "task-1316098" [ 1235.617792] env[61868]: _type = "Task" [ 1235.617792] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.625156] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316098, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.127009] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316098, 'name': ReconfigVM_Task, 'duration_secs': 0.127904} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1236.127336] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281640', 'volume_id': '622da5ad-c85c-484a-8297-2a235bd891eb', 'name': 'volume-622da5ad-c85c-484a-8297-2a235bd891eb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e', 'attached_at': '', 'detached_at': '', 'volume_id': '622da5ad-c85c-484a-8297-2a235bd891eb', 'serial': '622da5ad-c85c-484a-8297-2a235bd891eb'} {{(pid=61868) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1236.127873] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-96a32f44-c622-49d8-b6ba-f7b37d0644a8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.133302] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1236.133302] env[61868]: value = "task-1316099" [ 1236.133302] env[61868]: _type = "Task" [ 1236.133302] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1236.140369] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316099, 'name': Rename_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.643222] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316099, 'name': Rename_Task, 'duration_secs': 0.137325} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1236.643545] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1236.643826] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1ce0edf9-60dc-4729-af94-35ac5deb10ca {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.649618] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1236.649618] env[61868]: value = "task-1316100" [ 1236.649618] env[61868]: _type = "Task" [ 1236.649618] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1236.656847] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316100, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1237.159177] env[61868]: DEBUG oslo_vmware.api [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316100, 'name': PowerOnVM_Task, 'duration_secs': 0.452363} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1237.159552] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1237.159701] env[61868]: DEBUG nova.compute.manager [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1237.160640] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45f9f693-433a-4dc1-bc5b-6c0e2a3a5c0b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.676340] env[61868]: DEBUG oslo_concurrency.lockutils [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1237.676586] env[61868]: DEBUG oslo_concurrency.lockutils [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1237.677102] env[61868]: DEBUG nova.objects.instance [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61868) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1238.686866] env[61868]: DEBUG oslo_concurrency.lockutils [None req-80cf5d08-d419-49a1-9b26-ae4d548a37da tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1239.112292] env[61868]: DEBUG oslo_concurrency.lockutils [None req-025b2775-c9b5-443f-9dbb-a7bd602d78ec tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1239.112570] env[61868]: DEBUG oslo_concurrency.lockutils [None req-025b2775-c9b5-443f-9dbb-a7bd602d78ec tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1239.615524] env[61868]: INFO nova.compute.manager [None req-025b2775-c9b5-443f-9dbb-a7bd602d78ec tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Detaching volume 622da5ad-c85c-484a-8297-2a235bd891eb [ 1239.645491] env[61868]: INFO nova.virt.block_device [None req-025b2775-c9b5-443f-9dbb-a7bd602d78ec tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Attempting to driver detach volume 622da5ad-c85c-484a-8297-2a235bd891eb from mountpoint /dev/sdb [ 1239.645740] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-025b2775-c9b5-443f-9dbb-a7bd602d78ec tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Volume detach. Driver type: vmdk {{(pid=61868) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1239.645967] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-025b2775-c9b5-443f-9dbb-a7bd602d78ec tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281640', 'volume_id': '622da5ad-c85c-484a-8297-2a235bd891eb', 'name': 'volume-622da5ad-c85c-484a-8297-2a235bd891eb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e', 'attached_at': '', 'detached_at': '', 'volume_id': '622da5ad-c85c-484a-8297-2a235bd891eb', 'serial': '622da5ad-c85c-484a-8297-2a235bd891eb'} {{(pid=61868) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1239.646877] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40b881ab-d924-4235-83e7-da80576418ae {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.668557] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8668d950-4e7e-4cc5-9b96-a33fc02c9222 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.675168] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62dd603f-f82a-40bd-b005-db2bc39b04f6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.694574] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a5ee81a-bd40-4c46-9ea5-c039a961a0b9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.708673] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-025b2775-c9b5-443f-9dbb-a7bd602d78ec tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] The volume has not been displaced from its original location: [datastore1] volume-622da5ad-c85c-484a-8297-2a235bd891eb/volume-622da5ad-c85c-484a-8297-2a235bd891eb.vmdk. No consolidation needed. {{(pid=61868) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1239.713979] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-025b2775-c9b5-443f-9dbb-a7bd602d78ec tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Reconfiguring VM instance instance-0000006b to detach disk 2001 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1239.714279] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1fdfffc6-b654-4479-8a5b-cc1cd0cc7875 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.731839] env[61868]: DEBUG oslo_vmware.api [None req-025b2775-c9b5-443f-9dbb-a7bd602d78ec tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1239.731839] env[61868]: value = "task-1316101" [ 1239.731839] env[61868]: _type = "Task" [ 1239.731839] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1239.741017] env[61868]: DEBUG oslo_vmware.api [None req-025b2775-c9b5-443f-9dbb-a7bd602d78ec tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316101, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.241413] env[61868]: DEBUG oslo_vmware.api [None req-025b2775-c9b5-443f-9dbb-a7bd602d78ec tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316101, 'name': ReconfigVM_Task, 'duration_secs': 0.247271} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1240.241717] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-025b2775-c9b5-443f-9dbb-a7bd602d78ec tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Reconfigured VM instance instance-0000006b to detach disk 2001 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1240.246207] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9e98dc41-f2c9-473c-aa29-09ac1a3c7138 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.260271] env[61868]: DEBUG oslo_vmware.api [None req-025b2775-c9b5-443f-9dbb-a7bd602d78ec tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1240.260271] env[61868]: value = "task-1316102" [ 1240.260271] env[61868]: _type = "Task" [ 1240.260271] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1240.267279] env[61868]: DEBUG oslo_vmware.api [None req-025b2775-c9b5-443f-9dbb-a7bd602d78ec tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316102, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.769244] env[61868]: DEBUG oslo_vmware.api [None req-025b2775-c9b5-443f-9dbb-a7bd602d78ec tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316102, 'name': ReconfigVM_Task, 'duration_secs': 0.137982} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1240.769580] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-025b2775-c9b5-443f-9dbb-a7bd602d78ec tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281640', 'volume_id': '622da5ad-c85c-484a-8297-2a235bd891eb', 'name': 'volume-622da5ad-c85c-484a-8297-2a235bd891eb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e', 'attached_at': '', 'detached_at': '', 'volume_id': '622da5ad-c85c-484a-8297-2a235bd891eb', 'serial': '622da5ad-c85c-484a-8297-2a235bd891eb'} {{(pid=61868) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1241.316234] env[61868]: DEBUG nova.objects.instance [None req-025b2775-c9b5-443f-9dbb-a7bd602d78ec tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lazy-loading 'flavor' on Instance uuid 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1242.324077] env[61868]: DEBUG oslo_concurrency.lockutils [None req-025b2775-c9b5-443f-9dbb-a7bd602d78ec tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.211s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1243.338075] env[61868]: DEBUG oslo_concurrency.lockutils [None req-723c8d59-0a9d-4e98-a021-b00a3cdb3cae tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1243.338075] env[61868]: DEBUG oslo_concurrency.lockutils [None req-723c8d59-0a9d-4e98-a021-b00a3cdb3cae tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1243.338075] env[61868]: DEBUG oslo_concurrency.lockutils [None req-723c8d59-0a9d-4e98-a021-b00a3cdb3cae tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1243.338549] env[61868]: DEBUG oslo_concurrency.lockutils [None req-723c8d59-0a9d-4e98-a021-b00a3cdb3cae tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1243.338549] env[61868]: DEBUG oslo_concurrency.lockutils [None req-723c8d59-0a9d-4e98-a021-b00a3cdb3cae tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1243.340643] env[61868]: INFO nova.compute.manager [None req-723c8d59-0a9d-4e98-a021-b00a3cdb3cae tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Terminating instance [ 1243.342395] env[61868]: DEBUG nova.compute.manager [None req-723c8d59-0a9d-4e98-a021-b00a3cdb3cae tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1243.342583] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-723c8d59-0a9d-4e98-a021-b00a3cdb3cae tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1243.343418] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1739a83d-b15f-4702-a3e9-8da4f0505969 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.351668] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-723c8d59-0a9d-4e98-a021-b00a3cdb3cae tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1243.351911] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-283b82d7-93a0-4c23-a3fb-2acdeb3ad243 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.357826] env[61868]: DEBUG oslo_vmware.api [None req-723c8d59-0a9d-4e98-a021-b00a3cdb3cae tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1243.357826] env[61868]: value = "task-1316103" [ 1243.357826] env[61868]: _type = "Task" [ 1243.357826] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1243.365322] env[61868]: DEBUG oslo_vmware.api [None req-723c8d59-0a9d-4e98-a021-b00a3cdb3cae tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316103, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.867716] env[61868]: DEBUG oslo_vmware.api [None req-723c8d59-0a9d-4e98-a021-b00a3cdb3cae tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316103, 'name': PowerOffVM_Task, 'duration_secs': 0.186995} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1243.867986] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-723c8d59-0a9d-4e98-a021-b00a3cdb3cae tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1243.868188] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-723c8d59-0a9d-4e98-a021-b00a3cdb3cae tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1243.868443] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e35d8303-3c3b-4d3c-b73a-8d46b39c1acf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.929098] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-723c8d59-0a9d-4e98-a021-b00a3cdb3cae tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1243.929361] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-723c8d59-0a9d-4e98-a021-b00a3cdb3cae tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1243.929549] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-723c8d59-0a9d-4e98-a021-b00a3cdb3cae tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Deleting the datastore file [datastore1] 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1243.929817] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-db73fde4-9d35-4192-8ef7-3249c5119bb0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.937365] env[61868]: DEBUG oslo_vmware.api [None req-723c8d59-0a9d-4e98-a021-b00a3cdb3cae tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1243.937365] env[61868]: value = "task-1316105" [ 1243.937365] env[61868]: _type = "Task" [ 1243.937365] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1243.945202] env[61868]: DEBUG oslo_vmware.api [None req-723c8d59-0a9d-4e98-a021-b00a3cdb3cae tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316105, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.447228] env[61868]: DEBUG oslo_vmware.api [None req-723c8d59-0a9d-4e98-a021-b00a3cdb3cae tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316105, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13364} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.447658] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-723c8d59-0a9d-4e98-a021-b00a3cdb3cae tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1244.447658] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-723c8d59-0a9d-4e98-a021-b00a3cdb3cae tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1244.447828] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-723c8d59-0a9d-4e98-a021-b00a3cdb3cae tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1244.447998] env[61868]: INFO nova.compute.manager [None req-723c8d59-0a9d-4e98-a021-b00a3cdb3cae tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1244.448263] env[61868]: DEBUG oslo.service.loopingcall [None req-723c8d59-0a9d-4e98-a021-b00a3cdb3cae tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1244.448453] env[61868]: DEBUG nova.compute.manager [-] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1244.448546] env[61868]: DEBUG nova.network.neutron [-] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1244.890678] env[61868]: DEBUG nova.compute.manager [req-00abb1ef-719f-4580-806d-328749330739 req-dded05d6-08fb-4317-96ae-589adb3d27c3 service nova] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Received event network-vif-deleted-2bac9dd8-51e8-4aad-bf36-90e045aa9106 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1244.890865] env[61868]: INFO nova.compute.manager [req-00abb1ef-719f-4580-806d-328749330739 req-dded05d6-08fb-4317-96ae-589adb3d27c3 service nova] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Neutron deleted interface 2bac9dd8-51e8-4aad-bf36-90e045aa9106; detaching it from the instance and deleting it from the info cache [ 1244.891067] env[61868]: DEBUG nova.network.neutron [req-00abb1ef-719f-4580-806d-328749330739 req-dded05d6-08fb-4317-96ae-589adb3d27c3 service nova] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1245.363258] env[61868]: DEBUG nova.network.neutron [-] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1245.393432] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-353ca7fe-205f-4c30-b1f8-aa292669deda {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.403069] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b596dba-73eb-4e52-b623-1f7c451ade16 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.425739] env[61868]: DEBUG nova.compute.manager [req-00abb1ef-719f-4580-806d-328749330739 req-dded05d6-08fb-4317-96ae-589adb3d27c3 service nova] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Detach interface failed, port_id=2bac9dd8-51e8-4aad-bf36-90e045aa9106, reason: Instance 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e could not be found. {{(pid=61868) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1245.865659] env[61868]: INFO nova.compute.manager [-] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Took 1.42 seconds to deallocate network for instance. [ 1246.372142] env[61868]: DEBUG oslo_concurrency.lockutils [None req-723c8d59-0a9d-4e98-a021-b00a3cdb3cae tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1246.372425] env[61868]: DEBUG oslo_concurrency.lockutils [None req-723c8d59-0a9d-4e98-a021-b00a3cdb3cae tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1246.372650] env[61868]: DEBUG nova.objects.instance [None req-723c8d59-0a9d-4e98-a021-b00a3cdb3cae tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lazy-loading 'resources' on Instance uuid 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1246.926946] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-698d219c-1269-4852-b22f-d866d34f9659 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.934395] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-001eb980-6dae-4e40-838d-b778525f78a9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.964130] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad00aca0-a663-4c74-a654-9a77b30d71a7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.970777] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e398af75-038c-447c-bd34-8e89f217304e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.983282] env[61868]: DEBUG nova.compute.provider_tree [None req-723c8d59-0a9d-4e98-a021-b00a3cdb3cae tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1247.486933] env[61868]: DEBUG nova.scheduler.client.report [None req-723c8d59-0a9d-4e98-a021-b00a3cdb3cae tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1247.992492] env[61868]: DEBUG oslo_concurrency.lockutils [None req-723c8d59-0a9d-4e98-a021-b00a3cdb3cae tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.620s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1248.013183] env[61868]: INFO nova.scheduler.client.report [None req-723c8d59-0a9d-4e98-a021-b00a3cdb3cae tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Deleted allocations for instance 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e [ 1248.522517] env[61868]: DEBUG oslo_concurrency.lockutils [None req-723c8d59-0a9d-4e98-a021-b00a3cdb3cae tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.184s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1249.773550] env[61868]: DEBUG oslo_concurrency.lockutils [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "32abbff4-5d9f-4c4c-8270-e3e5d383825e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1249.773808] env[61868]: DEBUG oslo_concurrency.lockutils [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "32abbff4-5d9f-4c4c-8270-e3e5d383825e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1250.276366] env[61868]: DEBUG nova.compute.manager [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1250.797845] env[61868]: DEBUG oslo_concurrency.lockutils [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1250.798147] env[61868]: DEBUG oslo_concurrency.lockutils [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1250.799630] env[61868]: INFO nova.compute.claims [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1251.852314] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94606449-cf22-419f-8561-c686096f413e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.860994] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-540c3cdc-b37d-4fd6-b754-48f37d3fbfb7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.889575] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90501f8d-66ef-4ef8-8888-a5485d6a3b3c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.896257] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2525c4bc-de10-46fc-adf5-e8b606b97914 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.908863] env[61868]: DEBUG nova.compute.provider_tree [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1252.411971] env[61868]: DEBUG nova.scheduler.client.report [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1252.916808] env[61868]: DEBUG oslo_concurrency.lockutils [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.118s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1252.917378] env[61868]: DEBUG nova.compute.manager [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1253.422445] env[61868]: DEBUG nova.compute.utils [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1253.424012] env[61868]: DEBUG nova.compute.manager [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1253.424209] env[61868]: DEBUG nova.network.neutron [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1253.472402] env[61868]: DEBUG nova.policy [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '11a64f7a464b45c192edd10486c61138', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bdab2203834c4ee5bab11aca70e48cc1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 1253.723278] env[61868]: DEBUG nova.network.neutron [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Successfully created port: 6840da6b-2963-47e2-a5da-cf0e9d1d0fb8 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1253.927922] env[61868]: DEBUG nova.compute.manager [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1254.941041] env[61868]: DEBUG nova.compute.manager [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1254.965100] env[61868]: DEBUG nova.virt.hardware [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1254.965360] env[61868]: DEBUG nova.virt.hardware [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1254.965522] env[61868]: DEBUG nova.virt.hardware [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1254.965707] env[61868]: DEBUG nova.virt.hardware [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1254.965857] env[61868]: DEBUG nova.virt.hardware [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1254.966058] env[61868]: DEBUG nova.virt.hardware [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1254.966254] env[61868]: DEBUG nova.virt.hardware [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1254.966422] env[61868]: DEBUG nova.virt.hardware [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1254.966590] env[61868]: DEBUG nova.virt.hardware [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1254.966752] env[61868]: DEBUG nova.virt.hardware [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1254.966927] env[61868]: DEBUG nova.virt.hardware [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1254.967803] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51abea29-95c8-47fe-9a18-e93997deb4a8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.975684] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3726559c-9808-4be6-b981-96a7256bc42e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.116508] env[61868]: DEBUG nova.compute.manager [req-4e3b8c28-6d4e-4b22-bc59-bfe0f66c229a req-4a257d4f-301a-4b0b-80a6-87d103f08b5b service nova] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Received event network-vif-plugged-6840da6b-2963-47e2-a5da-cf0e9d1d0fb8 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1255.116782] env[61868]: DEBUG oslo_concurrency.lockutils [req-4e3b8c28-6d4e-4b22-bc59-bfe0f66c229a req-4a257d4f-301a-4b0b-80a6-87d103f08b5b service nova] Acquiring lock "32abbff4-5d9f-4c4c-8270-e3e5d383825e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1255.117033] env[61868]: DEBUG oslo_concurrency.lockutils [req-4e3b8c28-6d4e-4b22-bc59-bfe0f66c229a req-4a257d4f-301a-4b0b-80a6-87d103f08b5b service nova] Lock "32abbff4-5d9f-4c4c-8270-e3e5d383825e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1255.117356] env[61868]: DEBUG oslo_concurrency.lockutils [req-4e3b8c28-6d4e-4b22-bc59-bfe0f66c229a req-4a257d4f-301a-4b0b-80a6-87d103f08b5b service nova] Lock "32abbff4-5d9f-4c4c-8270-e3e5d383825e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1255.117486] env[61868]: DEBUG nova.compute.manager [req-4e3b8c28-6d4e-4b22-bc59-bfe0f66c229a req-4a257d4f-301a-4b0b-80a6-87d103f08b5b service nova] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] No waiting events found dispatching network-vif-plugged-6840da6b-2963-47e2-a5da-cf0e9d1d0fb8 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1255.117658] env[61868]: WARNING nova.compute.manager [req-4e3b8c28-6d4e-4b22-bc59-bfe0f66c229a req-4a257d4f-301a-4b0b-80a6-87d103f08b5b service nova] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Received unexpected event network-vif-plugged-6840da6b-2963-47e2-a5da-cf0e9d1d0fb8 for instance with vm_state building and task_state spawning. [ 1255.643448] env[61868]: DEBUG nova.network.neutron [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Successfully updated port: 6840da6b-2963-47e2-a5da-cf0e9d1d0fb8 {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1255.667342] env[61868]: DEBUG nova.compute.manager [req-499ad1db-a1c7-4a34-a2c1-f6b5e3beed0b req-33226f97-41c3-433e-b0f5-33cf417b25a5 service nova] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Received event network-changed-6840da6b-2963-47e2-a5da-cf0e9d1d0fb8 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1255.667477] env[61868]: DEBUG nova.compute.manager [req-499ad1db-a1c7-4a34-a2c1-f6b5e3beed0b req-33226f97-41c3-433e-b0f5-33cf417b25a5 service nova] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Refreshing instance network info cache due to event network-changed-6840da6b-2963-47e2-a5da-cf0e9d1d0fb8. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1255.667664] env[61868]: DEBUG oslo_concurrency.lockutils [req-499ad1db-a1c7-4a34-a2c1-f6b5e3beed0b req-33226f97-41c3-433e-b0f5-33cf417b25a5 service nova] Acquiring lock "refresh_cache-32abbff4-5d9f-4c4c-8270-e3e5d383825e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1255.667767] env[61868]: DEBUG oslo_concurrency.lockutils [req-499ad1db-a1c7-4a34-a2c1-f6b5e3beed0b req-33226f97-41c3-433e-b0f5-33cf417b25a5 service nova] Acquired lock "refresh_cache-32abbff4-5d9f-4c4c-8270-e3e5d383825e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1255.667972] env[61868]: DEBUG nova.network.neutron [req-499ad1db-a1c7-4a34-a2c1-f6b5e3beed0b req-33226f97-41c3-433e-b0f5-33cf417b25a5 service nova] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Refreshing network info cache for port 6840da6b-2963-47e2-a5da-cf0e9d1d0fb8 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1256.144695] env[61868]: DEBUG oslo_concurrency.lockutils [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "refresh_cache-32abbff4-5d9f-4c4c-8270-e3e5d383825e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1256.199463] env[61868]: DEBUG nova.network.neutron [req-499ad1db-a1c7-4a34-a2c1-f6b5e3beed0b req-33226f97-41c3-433e-b0f5-33cf417b25a5 service nova] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1256.268362] env[61868]: DEBUG nova.network.neutron [req-499ad1db-a1c7-4a34-a2c1-f6b5e3beed0b req-33226f97-41c3-433e-b0f5-33cf417b25a5 service nova] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1256.770507] env[61868]: DEBUG oslo_concurrency.lockutils [req-499ad1db-a1c7-4a34-a2c1-f6b5e3beed0b req-33226f97-41c3-433e-b0f5-33cf417b25a5 service nova] Releasing lock "refresh_cache-32abbff4-5d9f-4c4c-8270-e3e5d383825e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1256.770879] env[61868]: DEBUG oslo_concurrency.lockutils [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquired lock "refresh_cache-32abbff4-5d9f-4c4c-8270-e3e5d383825e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1256.771068] env[61868]: DEBUG nova.network.neutron [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1257.301968] env[61868]: DEBUG nova.network.neutron [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1257.420410] env[61868]: DEBUG nova.network.neutron [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Updating instance_info_cache with network_info: [{"id": "6840da6b-2963-47e2-a5da-cf0e9d1d0fb8", "address": "fa:16:3e:c4:fa:ea", "network": {"id": "16b3638c-1129-49b8-9f22-636ff7db7386", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1544179589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bdab2203834c4ee5bab11aca70e48cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d298db54-f13d-4bf6-b6c2-755074b3047f", "external-id": "nsx-vlan-transportzone-631", "segmentation_id": 631, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6840da6b-29", "ovs_interfaceid": "6840da6b-2963-47e2-a5da-cf0e9d1d0fb8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1257.923481] env[61868]: DEBUG oslo_concurrency.lockutils [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Releasing lock "refresh_cache-32abbff4-5d9f-4c4c-8270-e3e5d383825e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1257.923812] env[61868]: DEBUG nova.compute.manager [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Instance network_info: |[{"id": "6840da6b-2963-47e2-a5da-cf0e9d1d0fb8", "address": "fa:16:3e:c4:fa:ea", "network": {"id": "16b3638c-1129-49b8-9f22-636ff7db7386", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1544179589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bdab2203834c4ee5bab11aca70e48cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d298db54-f13d-4bf6-b6c2-755074b3047f", "external-id": "nsx-vlan-transportzone-631", "segmentation_id": 631, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6840da6b-29", "ovs_interfaceid": "6840da6b-2963-47e2-a5da-cf0e9d1d0fb8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1257.924291] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c4:fa:ea', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd298db54-f13d-4bf6-b6c2-755074b3047f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6840da6b-2963-47e2-a5da-cf0e9d1d0fb8', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1257.931681] env[61868]: DEBUG oslo.service.loopingcall [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1257.931896] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1257.932136] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-753c29c8-bead-4fef-94f4-b2401c978c18 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.952732] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1257.952732] env[61868]: value = "task-1316106" [ 1257.952732] env[61868]: _type = "Task" [ 1257.952732] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.959853] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1316106, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.462772] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1316106, 'name': CreateVM_Task, 'duration_secs': 0.299233} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.463135] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1258.463624] env[61868]: DEBUG oslo_concurrency.lockutils [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1258.463794] env[61868]: DEBUG oslo_concurrency.lockutils [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1258.464156] env[61868]: DEBUG oslo_concurrency.lockutils [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1258.464401] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e00fda36-0927-4bfa-8259-ccc37a72b6d5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.468588] env[61868]: DEBUG oslo_vmware.api [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1258.468588] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52af2712-e389-88e2-7b69-267ce464981d" [ 1258.468588] env[61868]: _type = "Task" [ 1258.468588] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.476905] env[61868]: DEBUG oslo_vmware.api [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52af2712-e389-88e2-7b69-267ce464981d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.978190] env[61868]: DEBUG oslo_vmware.api [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52af2712-e389-88e2-7b69-267ce464981d, 'name': SearchDatastore_Task, 'duration_secs': 0.009014} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.978492] env[61868]: DEBUG oslo_concurrency.lockutils [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1258.978726] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1258.978963] env[61868]: DEBUG oslo_concurrency.lockutils [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1258.979129] env[61868]: DEBUG oslo_concurrency.lockutils [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1258.979317] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1258.979568] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5aefe3ed-920f-4fd6-8364-141e4557647e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.987119] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1258.987294] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1258.987963] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ec5e6dbd-0c31-4cfe-9dbd-267a1811b3d0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.992535] env[61868]: DEBUG oslo_vmware.api [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1258.992535] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52fdb38b-49ed-c4e2-d80a-94f06263a3da" [ 1258.992535] env[61868]: _type = "Task" [ 1258.992535] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.999373] env[61868]: DEBUG oslo_vmware.api [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52fdb38b-49ed-c4e2-d80a-94f06263a3da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.502046] env[61868]: DEBUG oslo_vmware.api [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52fdb38b-49ed-c4e2-d80a-94f06263a3da, 'name': SearchDatastore_Task, 'duration_secs': 0.008036} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1259.503795] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-29cc0882-c4a7-4417-8b55-95ea5dbbd61c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.507953] env[61868]: DEBUG oslo_vmware.api [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1259.507953] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5242b677-985e-43e5-9ade-ce57bee1e37a" [ 1259.507953] env[61868]: _type = "Task" [ 1259.507953] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1259.514861] env[61868]: DEBUG oslo_vmware.api [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5242b677-985e-43e5-9ade-ce57bee1e37a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.018472] env[61868]: DEBUG oslo_vmware.api [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5242b677-985e-43e5-9ade-ce57bee1e37a, 'name': SearchDatastore_Task, 'duration_secs': 0.008673} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1260.018737] env[61868]: DEBUG oslo_concurrency.lockutils [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1260.019064] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 32abbff4-5d9f-4c4c-8270-e3e5d383825e/32abbff4-5d9f-4c4c-8270-e3e5d383825e.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1260.019301] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-11379004-b8d9-46fd-9b25-8d31915c4223 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.026418] env[61868]: DEBUG oslo_vmware.api [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1260.026418] env[61868]: value = "task-1316107" [ 1260.026418] env[61868]: _type = "Task" [ 1260.026418] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1260.033580] env[61868]: DEBUG oslo_vmware.api [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316107, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.536582] env[61868]: DEBUG oslo_vmware.api [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316107, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.504072} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1260.536936] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 32abbff4-5d9f-4c4c-8270-e3e5d383825e/32abbff4-5d9f-4c4c-8270-e3e5d383825e.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1260.537083] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1260.537339] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-20f6ba8f-d651-4649-aeae-6d2ffa784d33 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.543590] env[61868]: DEBUG oslo_vmware.api [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1260.543590] env[61868]: value = "task-1316108" [ 1260.543590] env[61868]: _type = "Task" [ 1260.543590] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1260.552812] env[61868]: DEBUG oslo_vmware.api [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316108, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.053623] env[61868]: DEBUG oslo_vmware.api [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316108, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065819} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1261.053892] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1261.054669] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-288ec1a9-c87e-488d-8397-e8f89af8a136 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.075750] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] 32abbff4-5d9f-4c4c-8270-e3e5d383825e/32abbff4-5d9f-4c4c-8270-e3e5d383825e.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1261.076007] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-31bdb3f7-d4c4-4868-8019-c89dae47260a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.093939] env[61868]: DEBUG oslo_vmware.api [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1261.093939] env[61868]: value = "task-1316109" [ 1261.093939] env[61868]: _type = "Task" [ 1261.093939] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1261.101110] env[61868]: DEBUG oslo_vmware.api [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316109, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.603519] env[61868]: DEBUG oslo_vmware.api [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316109, 'name': ReconfigVM_Task, 'duration_secs': 0.306673} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1261.604015] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Reconfigured VM instance instance-0000006d to attach disk [datastore1] 32abbff4-5d9f-4c4c-8270-e3e5d383825e/32abbff4-5d9f-4c4c-8270-e3e5d383825e.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1261.604466] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-92e32fbe-22db-494c-a9b0-4b8e20284b44 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.610266] env[61868]: DEBUG oslo_vmware.api [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1261.610266] env[61868]: value = "task-1316110" [ 1261.610266] env[61868]: _type = "Task" [ 1261.610266] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1261.617930] env[61868]: DEBUG oslo_vmware.api [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316110, 'name': Rename_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.120554] env[61868]: DEBUG oslo_vmware.api [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316110, 'name': Rename_Task, 'duration_secs': 0.139651} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1262.120855] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1262.121115] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e3c7c4b4-770d-4a7b-b1d3-d95e36534ed7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.127383] env[61868]: DEBUG oslo_vmware.api [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1262.127383] env[61868]: value = "task-1316111" [ 1262.127383] env[61868]: _type = "Task" [ 1262.127383] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1262.136065] env[61868]: DEBUG oslo_vmware.api [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316111, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.636158] env[61868]: DEBUG oslo_vmware.api [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316111, 'name': PowerOnVM_Task, 'duration_secs': 0.487964} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1262.636523] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1262.638078] env[61868]: INFO nova.compute.manager [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Took 7.70 seconds to spawn the instance on the hypervisor. [ 1262.638078] env[61868]: DEBUG nova.compute.manager [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1262.638078] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27b1baed-144b-46ca-bf23-3f74494b655c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.154826] env[61868]: INFO nova.compute.manager [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Took 12.37 seconds to build instance. [ 1263.657455] env[61868]: DEBUG oslo_concurrency.lockutils [None req-50b4085d-f0e0-432e-acd9-ba8139a1a65c tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "32abbff4-5d9f-4c4c-8270-e3e5d383825e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.883s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1263.665585] env[61868]: DEBUG nova.compute.manager [req-8f51fd9c-b50c-4a12-8578-015fff402ab4 req-99a63181-fa6c-4dfd-81cb-d22fdda09395 service nova] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Received event network-changed-6840da6b-2963-47e2-a5da-cf0e9d1d0fb8 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1263.665718] env[61868]: DEBUG nova.compute.manager [req-8f51fd9c-b50c-4a12-8578-015fff402ab4 req-99a63181-fa6c-4dfd-81cb-d22fdda09395 service nova] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Refreshing instance network info cache due to event network-changed-6840da6b-2963-47e2-a5da-cf0e9d1d0fb8. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1263.666305] env[61868]: DEBUG oslo_concurrency.lockutils [req-8f51fd9c-b50c-4a12-8578-015fff402ab4 req-99a63181-fa6c-4dfd-81cb-d22fdda09395 service nova] Acquiring lock "refresh_cache-32abbff4-5d9f-4c4c-8270-e3e5d383825e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1263.666305] env[61868]: DEBUG oslo_concurrency.lockutils [req-8f51fd9c-b50c-4a12-8578-015fff402ab4 req-99a63181-fa6c-4dfd-81cb-d22fdda09395 service nova] Acquired lock "refresh_cache-32abbff4-5d9f-4c4c-8270-e3e5d383825e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1263.666434] env[61868]: DEBUG nova.network.neutron [req-8f51fd9c-b50c-4a12-8578-015fff402ab4 req-99a63181-fa6c-4dfd-81cb-d22fdda09395 service nova] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Refreshing network info cache for port 6840da6b-2963-47e2-a5da-cf0e9d1d0fb8 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1264.363861] env[61868]: DEBUG nova.network.neutron [req-8f51fd9c-b50c-4a12-8578-015fff402ab4 req-99a63181-fa6c-4dfd-81cb-d22fdda09395 service nova] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Updated VIF entry in instance network info cache for port 6840da6b-2963-47e2-a5da-cf0e9d1d0fb8. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1264.364280] env[61868]: DEBUG nova.network.neutron [req-8f51fd9c-b50c-4a12-8578-015fff402ab4 req-99a63181-fa6c-4dfd-81cb-d22fdda09395 service nova] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Updating instance_info_cache with network_info: [{"id": "6840da6b-2963-47e2-a5da-cf0e9d1d0fb8", "address": "fa:16:3e:c4:fa:ea", "network": {"id": "16b3638c-1129-49b8-9f22-636ff7db7386", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1544179589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bdab2203834c4ee5bab11aca70e48cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d298db54-f13d-4bf6-b6c2-755074b3047f", "external-id": "nsx-vlan-transportzone-631", "segmentation_id": 631, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6840da6b-29", "ovs_interfaceid": "6840da6b-2963-47e2-a5da-cf0e9d1d0fb8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1264.867035] env[61868]: DEBUG oslo_concurrency.lockutils [req-8f51fd9c-b50c-4a12-8578-015fff402ab4 req-99a63181-fa6c-4dfd-81cb-d22fdda09395 service nova] Releasing lock "refresh_cache-32abbff4-5d9f-4c4c-8270-e3e5d383825e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1268.091606] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0d1d800a-1860-4164-a412-704fa6aff08e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "e18dfd04-5a37-44d1-930b-52b50362755b" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1268.091892] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0d1d800a-1860-4164-a412-704fa6aff08e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "e18dfd04-5a37-44d1-930b-52b50362755b" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1268.595127] env[61868]: DEBUG nova.compute.utils [None req-0d1d800a-1860-4164-a412-704fa6aff08e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1269.098416] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0d1d800a-1860-4164-a412-704fa6aff08e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "e18dfd04-5a37-44d1-930b-52b50362755b" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1270.156305] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0d1d800a-1860-4164-a412-704fa6aff08e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "e18dfd04-5a37-44d1-930b-52b50362755b" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1270.156699] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0d1d800a-1860-4164-a412-704fa6aff08e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "e18dfd04-5a37-44d1-930b-52b50362755b" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1270.156891] env[61868]: INFO nova.compute.manager [None req-0d1d800a-1860-4164-a412-704fa6aff08e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Attaching volume 6bb58fe7-5636-40f7-8375-5dd71b90fca3 to /dev/sdb [ 1270.188042] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94c7bd0d-5fab-4d72-ae65-3ece738673d1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.196399] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a923f80-7573-4076-8251-82179b0eb81e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.210736] env[61868]: DEBUG nova.virt.block_device [None req-0d1d800a-1860-4164-a412-704fa6aff08e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Updating existing volume attachment record: 83964cf0-8dfc-49b3-9f9e-90f7e2e91073 {{(pid=61868) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1274.754125] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d1d800a-1860-4164-a412-704fa6aff08e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Volume attach. Driver type: vmdk {{(pid=61868) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1274.754382] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d1d800a-1860-4164-a412-704fa6aff08e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281646', 'volume_id': '6bb58fe7-5636-40f7-8375-5dd71b90fca3', 'name': 'volume-6bb58fe7-5636-40f7-8375-5dd71b90fca3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e18dfd04-5a37-44d1-930b-52b50362755b', 'attached_at': '', 'detached_at': '', 'volume_id': '6bb58fe7-5636-40f7-8375-5dd71b90fca3', 'serial': '6bb58fe7-5636-40f7-8375-5dd71b90fca3'} {{(pid=61868) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1274.755278] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49920588-3056-4315-b3a9-a0861ceda23f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.771256] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c87fdb31-6f19-49a0-82fd-42ea37c49dcb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.794318] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d1d800a-1860-4164-a412-704fa6aff08e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] volume-6bb58fe7-5636-40f7-8375-5dd71b90fca3/volume-6bb58fe7-5636-40f7-8375-5dd71b90fca3.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1274.794561] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-30df274a-a6ad-4e51-af5e-25eddbc109f7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.812336] env[61868]: DEBUG oslo_vmware.api [None req-0d1d800a-1860-4164-a412-704fa6aff08e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1274.812336] env[61868]: value = "task-1316121" [ 1274.812336] env[61868]: _type = "Task" [ 1274.812336] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1274.820077] env[61868]: DEBUG oslo_vmware.api [None req-0d1d800a-1860-4164-a412-704fa6aff08e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316121, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1275.075213] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1275.075404] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1275.075560] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61868) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1275.322387] env[61868]: DEBUG oslo_vmware.api [None req-0d1d800a-1860-4164-a412-704fa6aff08e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316121, 'name': ReconfigVM_Task, 'duration_secs': 0.38059} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1275.322641] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d1d800a-1860-4164-a412-704fa6aff08e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Reconfigured VM instance instance-0000006c to attach disk [datastore1] volume-6bb58fe7-5636-40f7-8375-5dd71b90fca3/volume-6bb58fe7-5636-40f7-8375-5dd71b90fca3.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1275.327306] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6239c8c7-385f-4093-8596-5f83f59af1fa {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.342770] env[61868]: DEBUG oslo_vmware.api [None req-0d1d800a-1860-4164-a412-704fa6aff08e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1275.342770] env[61868]: value = "task-1316122" [ 1275.342770] env[61868]: _type = "Task" [ 1275.342770] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1275.350375] env[61868]: DEBUG oslo_vmware.api [None req-0d1d800a-1860-4164-a412-704fa6aff08e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316122, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1275.853700] env[61868]: DEBUG oslo_vmware.api [None req-0d1d800a-1860-4164-a412-704fa6aff08e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316122, 'name': ReconfigVM_Task, 'duration_secs': 0.144568} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1275.854049] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d1d800a-1860-4164-a412-704fa6aff08e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281646', 'volume_id': '6bb58fe7-5636-40f7-8375-5dd71b90fca3', 'name': 'volume-6bb58fe7-5636-40f7-8375-5dd71b90fca3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e18dfd04-5a37-44d1-930b-52b50362755b', 'attached_at': '', 'detached_at': '', 'volume_id': '6bb58fe7-5636-40f7-8375-5dd71b90fca3', 'serial': '6bb58fe7-5636-40f7-8375-5dd71b90fca3'} {{(pid=61868) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1276.889867] env[61868]: DEBUG nova.objects.instance [None req-0d1d800a-1860-4164-a412-704fa6aff08e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lazy-loading 'flavor' on Instance uuid e18dfd04-5a37-44d1-930b-52b50362755b {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1277.395015] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0d1d800a-1860-4164-a412-704fa6aff08e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "e18dfd04-5a37-44d1-930b-52b50362755b" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.238s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1278.075390] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1278.075390] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager.update_available_resource {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1278.303847] env[61868]: DEBUG nova.compute.manager [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Stashing vm_state: active {{(pid=61868) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1278.578556] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1278.578794] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1278.578973] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1278.579143] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61868) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1278.580055] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13e85842-a776-41ae-96f2-14d2b7758734 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.589401] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40fcb82d-cf22-4698-a01f-3e6865534a7b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.603237] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d526523-b890-40f5-8264-c15ffeca1864 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.609516] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3664564-a353-419c-aa16-c4dc8e9e9b66 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.637725] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181123MB free_disk=134GB free_vcpus=48 pci_devices=None {{(pid=61868) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1278.637864] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1278.638063] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1278.820646] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1279.644995] env[61868]: INFO nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Updating resource usage from migration 0cad9d16-b904-461e-8bbb-0cb210983e12 [ 1279.662361] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 59171bf1-f2d4-4ff5-a661-025f5dca7600 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1279.662532] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 32abbff4-5d9f-4c4c-8270-e3e5d383825e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1279.662659] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Migration 0cad9d16-b904-461e-8bbb-0cb210983e12 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1279.662778] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance e18dfd04-5a37-44d1-930b-52b50362755b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1279.662956] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=61868) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1279.663106] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=61868) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1279.719949] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03ac65fb-d80e-4c69-9eab-1418957c4272 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.727128] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5c55fcc-bb13-41f5-99c1-a849c16ae9d1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.756378] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7651ff10-f06d-4e9f-874b-e40452b8a714 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.763929] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ec85b71-dc50-4ccf-aad0-9e11795b4a5d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.777284] env[61868]: DEBUG nova.compute.provider_tree [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1280.280801] env[61868]: DEBUG nova.scheduler.client.report [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1280.786102] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61868) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1280.786490] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.148s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1280.786641] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 1.966s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1281.291640] env[61868]: INFO nova.compute.claims [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1281.783260] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1281.783497] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1281.783652] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Starting heal instance info cache {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1281.783763] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Rebuilding the list of instances to heal {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1281.798313] env[61868]: INFO nova.compute.resource_tracker [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Updating resource usage from migration 0cad9d16-b904-461e-8bbb-0cb210983e12 [ 1281.852362] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d6340b7-dcb3-42c4-bf6f-7879eea80ae4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.860084] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ba726e0-c4f5-40e3-b0c5-fcda930898db {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.888848] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25cbe68a-24c2-4cda-b96b-78f6e6d8f5fe {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.896081] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-024899b3-9dd4-4a18-a0a0-43df4bb36509 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.909623] env[61868]: DEBUG nova.compute.provider_tree [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1282.325395] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "refresh_cache-59171bf1-f2d4-4ff5-a661-025f5dca7600" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1282.325552] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquired lock "refresh_cache-59171bf1-f2d4-4ff5-a661-025f5dca7600" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1282.325694] env[61868]: DEBUG nova.network.neutron [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Forcefully refreshing network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1282.325847] env[61868]: DEBUG nova.objects.instance [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lazy-loading 'info_cache' on Instance uuid 59171bf1-f2d4-4ff5-a661-025f5dca7600 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1282.412722] env[61868]: DEBUG nova.scheduler.client.report [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1282.873671] env[61868]: DEBUG oslo_concurrency.lockutils [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "77e19bcd-5963-4c48-bcf2-da5c92b142ff" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1282.873671] env[61868]: DEBUG oslo_concurrency.lockutils [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "77e19bcd-5963-4c48-bcf2-da5c92b142ff" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1282.917186] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.130s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1282.917413] env[61868]: INFO nova.compute.manager [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Migrating [ 1283.375236] env[61868]: DEBUG nova.compute.manager [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1283.431056] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "refresh_cache-e18dfd04-5a37-44d1-930b-52b50362755b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1283.431318] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquired lock "refresh_cache-e18dfd04-5a37-44d1-930b-52b50362755b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1283.431493] env[61868]: DEBUG nova.network.neutron [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1283.897919] env[61868]: DEBUG oslo_concurrency.lockutils [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1283.898211] env[61868]: DEBUG oslo_concurrency.lockutils [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1283.899694] env[61868]: INFO nova.compute.claims [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1284.048781] env[61868]: DEBUG nova.network.neutron [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Updating instance_info_cache with network_info: [{"id": "8f9ad20f-f367-4826-8aae-00a1192d0bc8", "address": "fa:16:3e:63:2a:68", "network": {"id": "16b3638c-1129-49b8-9f22-636ff7db7386", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1544179589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.190", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bdab2203834c4ee5bab11aca70e48cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d298db54-f13d-4bf6-b6c2-755074b3047f", "external-id": "nsx-vlan-transportzone-631", "segmentation_id": 631, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f9ad20f-f3", "ovs_interfaceid": "8f9ad20f-f367-4826-8aae-00a1192d0bc8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1284.127604] env[61868]: DEBUG nova.network.neutron [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Updating instance_info_cache with network_info: [{"id": "644d55b3-8c23-4a45-93d6-9d5464d1631e", "address": "fa:16:3e:db:75:9d", "network": {"id": "7e8cbecf-dfdd-43f2-8745-ae1d2d55d309", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-33935587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a7014141b24101ba7920bc83be4a9e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4db2961d-273d-4634-9d06-a94fa9d384fb", "external-id": "nsx-vlan-transportzone-572", "segmentation_id": 572, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap644d55b3-8c", "ovs_interfaceid": "644d55b3-8c23-4a45-93d6-9d5464d1631e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1284.552020] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Releasing lock "refresh_cache-59171bf1-f2d4-4ff5-a661-025f5dca7600" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1284.552250] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Updated the network info_cache for instance {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1284.552455] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1284.552620] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1284.552782] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1284.629772] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Releasing lock "refresh_cache-e18dfd04-5a37-44d1-930b-52b50362755b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1284.975230] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0cfbb78-5439-46fa-8738-5e995305e8c5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.983716] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf3706ce-920a-4935-864b-c797ab9dd69b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.013600] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eca6612a-e2a1-471d-8205-5e946b4b8819 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.021192] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f6c3bfc-554f-4bdf-b481-051f6c276338 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.034223] env[61868]: DEBUG nova.compute.provider_tree [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1285.537926] env[61868]: DEBUG nova.scheduler.client.report [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1286.042624] env[61868]: DEBUG oslo_concurrency.lockutils [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.144s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1286.043018] env[61868]: DEBUG nova.compute.manager [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1286.143466] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47a26e7f-86a0-4a53-978d-6d786b69de03 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.164594] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Updating instance 'e18dfd04-5a37-44d1-930b-52b50362755b' progress to 0 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1286.548501] env[61868]: DEBUG nova.compute.utils [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1286.551276] env[61868]: DEBUG nova.compute.manager [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1286.551459] env[61868]: DEBUG nova.network.neutron [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1286.588194] env[61868]: DEBUG nova.policy [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '11a64f7a464b45c192edd10486c61138', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bdab2203834c4ee5bab11aca70e48cc1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 1286.670330] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1286.670645] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3d5f0cae-71a3-4447-82b6-1381d6444315 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.680636] env[61868]: DEBUG oslo_vmware.api [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1286.680636] env[61868]: value = "task-1316126" [ 1286.680636] env[61868]: _type = "Task" [ 1286.680636] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1286.690128] env[61868]: DEBUG oslo_vmware.api [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316126, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1286.877643] env[61868]: DEBUG nova.network.neutron [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Successfully created port: eab3dce4-b36e-4ab7-bc96-27545c10e311 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1287.055763] env[61868]: DEBUG nova.compute.manager [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1287.191670] env[61868]: DEBUG oslo_vmware.api [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316126, 'name': PowerOffVM_Task, 'duration_secs': 0.235032} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1287.191961] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1287.192175] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Updating instance 'e18dfd04-5a37-44d1-930b-52b50362755b' progress to 17 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1287.562056] env[61868]: INFO nova.virt.block_device [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Booting with volume ec4ed682-190b-4dbe-b2f8-1b4e3e82fb91 at /dev/sda [ 1287.601051] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8460fd02-e4de-425a-8d72-238d38082968 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.611094] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9d4f46f-1fcb-4bc7-8e33-ce3bb39d76e9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.638355] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7647996a-e125-4716-b98f-45ae2ae79b61 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.648377] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82b202eb-8fe1-4708-aa8c-4181a23e2466 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.674547] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d5ad20d-0d9c-4f25-978e-5ff12d3fc2b1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.682148] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11be3c01-125d-48bc-810b-cd8a13693873 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.699077] env[61868]: DEBUG nova.virt.hardware [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1287.699359] env[61868]: DEBUG nova.virt.hardware [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1287.699487] env[61868]: DEBUG nova.virt.hardware [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1287.699674] env[61868]: DEBUG nova.virt.hardware [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1287.699823] env[61868]: DEBUG nova.virt.hardware [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1287.700047] env[61868]: DEBUG nova.virt.hardware [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1287.700202] env[61868]: DEBUG nova.virt.hardware [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1287.700369] env[61868]: DEBUG nova.virt.hardware [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1287.700540] env[61868]: DEBUG nova.virt.hardware [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1287.700711] env[61868]: DEBUG nova.virt.hardware [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1287.701021] env[61868]: DEBUG nova.virt.hardware [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1287.705890] env[61868]: DEBUG nova.virt.block_device [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Updating existing volume attachment record: 8b042f35-f3bf-454e-b880-851893ecde7c {{(pid=61868) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1287.708335] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e302c7d6-e4cc-4619-aa49-15feefc0336a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.725130] env[61868]: DEBUG oslo_vmware.api [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1287.725130] env[61868]: value = "task-1316127" [ 1287.725130] env[61868]: _type = "Task" [ 1287.725130] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1287.736054] env[61868]: DEBUG oslo_vmware.api [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316127, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1288.236887] env[61868]: DEBUG oslo_vmware.api [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316127, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1288.394785] env[61868]: DEBUG nova.compute.manager [req-c1570be5-5950-4821-a845-1a34e9e01eff req-3f8dc5cf-891b-4063-a6a2-d7a1539478f1 service nova] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Received event network-vif-plugged-eab3dce4-b36e-4ab7-bc96-27545c10e311 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1288.395391] env[61868]: DEBUG oslo_concurrency.lockutils [req-c1570be5-5950-4821-a845-1a34e9e01eff req-3f8dc5cf-891b-4063-a6a2-d7a1539478f1 service nova] Acquiring lock "77e19bcd-5963-4c48-bcf2-da5c92b142ff-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1288.395590] env[61868]: DEBUG oslo_concurrency.lockutils [req-c1570be5-5950-4821-a845-1a34e9e01eff req-3f8dc5cf-891b-4063-a6a2-d7a1539478f1 service nova] Lock "77e19bcd-5963-4c48-bcf2-da5c92b142ff-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1288.395767] env[61868]: DEBUG oslo_concurrency.lockutils [req-c1570be5-5950-4821-a845-1a34e9e01eff req-3f8dc5cf-891b-4063-a6a2-d7a1539478f1 service nova] Lock "77e19bcd-5963-4c48-bcf2-da5c92b142ff-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1288.395939] env[61868]: DEBUG nova.compute.manager [req-c1570be5-5950-4821-a845-1a34e9e01eff req-3f8dc5cf-891b-4063-a6a2-d7a1539478f1 service nova] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] No waiting events found dispatching network-vif-plugged-eab3dce4-b36e-4ab7-bc96-27545c10e311 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1288.396120] env[61868]: WARNING nova.compute.manager [req-c1570be5-5950-4821-a845-1a34e9e01eff req-3f8dc5cf-891b-4063-a6a2-d7a1539478f1 service nova] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Received unexpected event network-vif-plugged-eab3dce4-b36e-4ab7-bc96-27545c10e311 for instance with vm_state building and task_state block_device_mapping. [ 1288.476517] env[61868]: DEBUG nova.network.neutron [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Successfully updated port: eab3dce4-b36e-4ab7-bc96-27545c10e311 {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1288.737081] env[61868]: DEBUG oslo_vmware.api [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316127, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1288.978862] env[61868]: DEBUG oslo_concurrency.lockutils [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "refresh_cache-77e19bcd-5963-4c48-bcf2-da5c92b142ff" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1288.979053] env[61868]: DEBUG oslo_concurrency.lockutils [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquired lock "refresh_cache-77e19bcd-5963-4c48-bcf2-da5c92b142ff" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1288.979408] env[61868]: DEBUG nova.network.neutron [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1289.236686] env[61868]: DEBUG oslo_vmware.api [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316127, 'name': ReconfigVM_Task, 'duration_secs': 1.175329} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1289.236945] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Updating instance 'e18dfd04-5a37-44d1-930b-52b50362755b' progress to 33 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1289.507542] env[61868]: DEBUG nova.network.neutron [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1289.624519] env[61868]: DEBUG nova.network.neutron [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Updating instance_info_cache with network_info: [{"id": "eab3dce4-b36e-4ab7-bc96-27545c10e311", "address": "fa:16:3e:4a:7e:64", "network": {"id": "16b3638c-1129-49b8-9f22-636ff7db7386", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1544179589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bdab2203834c4ee5bab11aca70e48cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d298db54-f13d-4bf6-b6c2-755074b3047f", "external-id": "nsx-vlan-transportzone-631", "segmentation_id": 631, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeab3dce4-b3", "ovs_interfaceid": "eab3dce4-b36e-4ab7-bc96-27545c10e311", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1289.742805] env[61868]: DEBUG nova.virt.hardware [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1289.743095] env[61868]: DEBUG nova.virt.hardware [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1289.743243] env[61868]: DEBUG nova.virt.hardware [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1289.743428] env[61868]: DEBUG nova.virt.hardware [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1289.743580] env[61868]: DEBUG nova.virt.hardware [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1289.743734] env[61868]: DEBUG nova.virt.hardware [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1289.743942] env[61868]: DEBUG nova.virt.hardware [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1289.744118] env[61868]: DEBUG nova.virt.hardware [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1289.744290] env[61868]: DEBUG nova.virt.hardware [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1289.744458] env[61868]: DEBUG nova.virt.hardware [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1289.744634] env[61868]: DEBUG nova.virt.hardware [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1289.750054] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Reconfiguring VM instance instance-0000006c to detach disk 2000 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1289.750344] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7a715b8d-3032-4e53-936b-b8d69aeff8ce {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.769968] env[61868]: DEBUG oslo_vmware.api [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1289.769968] env[61868]: value = "task-1316128" [ 1289.769968] env[61868]: _type = "Task" [ 1289.769968] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1289.780380] env[61868]: DEBUG oslo_vmware.api [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316128, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1289.789836] env[61868]: DEBUG nova.compute.manager [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1289.790381] env[61868]: DEBUG nova.virt.hardware [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1289.790638] env[61868]: DEBUG nova.virt.hardware [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1289.790824] env[61868]: DEBUG nova.virt.hardware [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1289.791017] env[61868]: DEBUG nova.virt.hardware [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1289.791175] env[61868]: DEBUG nova.virt.hardware [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1289.791325] env[61868]: DEBUG nova.virt.hardware [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1289.791530] env[61868]: DEBUG nova.virt.hardware [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1289.791694] env[61868]: DEBUG nova.virt.hardware [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1289.791896] env[61868]: DEBUG nova.virt.hardware [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1289.792029] env[61868]: DEBUG nova.virt.hardware [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1289.792212] env[61868]: DEBUG nova.virt.hardware [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1289.793017] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fcd49a4-95c3-46bd-a1e3-da5dd2b21ab9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.800981] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f99165a-7075-4e3a-b414-42c961a0e9d3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.127470] env[61868]: DEBUG oslo_concurrency.lockutils [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Releasing lock "refresh_cache-77e19bcd-5963-4c48-bcf2-da5c92b142ff" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1290.127860] env[61868]: DEBUG nova.compute.manager [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Instance network_info: |[{"id": "eab3dce4-b36e-4ab7-bc96-27545c10e311", "address": "fa:16:3e:4a:7e:64", "network": {"id": "16b3638c-1129-49b8-9f22-636ff7db7386", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1544179589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bdab2203834c4ee5bab11aca70e48cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d298db54-f13d-4bf6-b6c2-755074b3047f", "external-id": "nsx-vlan-transportzone-631", "segmentation_id": 631, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeab3dce4-b3", "ovs_interfaceid": "eab3dce4-b36e-4ab7-bc96-27545c10e311", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1290.128332] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4a:7e:64', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd298db54-f13d-4bf6-b6c2-755074b3047f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'eab3dce4-b36e-4ab7-bc96-27545c10e311', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1290.135792] env[61868]: DEBUG oslo.service.loopingcall [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1290.136013] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1290.136239] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-be62f9ef-7db4-482a-9b8c-b220a2e2c572 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.156527] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1290.156527] env[61868]: value = "task-1316129" [ 1290.156527] env[61868]: _type = "Task" [ 1290.156527] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1290.164163] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1316129, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1290.280282] env[61868]: DEBUG oslo_vmware.api [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316128, 'name': ReconfigVM_Task, 'duration_secs': 0.211841} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1290.280704] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Reconfigured VM instance instance-0000006c to detach disk 2000 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1290.281381] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b5b0143-3929-48a1-a3f4-3a5f1c61a3e6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.306617] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] e18dfd04-5a37-44d1-930b-52b50362755b/e18dfd04-5a37-44d1-930b-52b50362755b.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1290.306909] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-858af3f3-15e4-4a36-835a-c236e726bba2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.325269] env[61868]: DEBUG oslo_vmware.api [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1290.325269] env[61868]: value = "task-1316130" [ 1290.325269] env[61868]: _type = "Task" [ 1290.325269] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1290.333306] env[61868]: DEBUG oslo_vmware.api [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316130, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1290.423361] env[61868]: DEBUG nova.compute.manager [req-bd75bc31-7188-4495-a3d6-18e72dc672bd req-b52dd208-d637-4286-8d02-7b4992961a3b service nova] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Received event network-changed-eab3dce4-b36e-4ab7-bc96-27545c10e311 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1290.423361] env[61868]: DEBUG nova.compute.manager [req-bd75bc31-7188-4495-a3d6-18e72dc672bd req-b52dd208-d637-4286-8d02-7b4992961a3b service nova] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Refreshing instance network info cache due to event network-changed-eab3dce4-b36e-4ab7-bc96-27545c10e311. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1290.423361] env[61868]: DEBUG oslo_concurrency.lockutils [req-bd75bc31-7188-4495-a3d6-18e72dc672bd req-b52dd208-d637-4286-8d02-7b4992961a3b service nova] Acquiring lock "refresh_cache-77e19bcd-5963-4c48-bcf2-da5c92b142ff" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1290.423538] env[61868]: DEBUG oslo_concurrency.lockutils [req-bd75bc31-7188-4495-a3d6-18e72dc672bd req-b52dd208-d637-4286-8d02-7b4992961a3b service nova] Acquired lock "refresh_cache-77e19bcd-5963-4c48-bcf2-da5c92b142ff" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1290.423627] env[61868]: DEBUG nova.network.neutron [req-bd75bc31-7188-4495-a3d6-18e72dc672bd req-b52dd208-d637-4286-8d02-7b4992961a3b service nova] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Refreshing network info cache for port eab3dce4-b36e-4ab7-bc96-27545c10e311 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1290.667146] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1316129, 'name': CreateVM_Task, 'duration_secs': 0.295408} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1290.667344] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1290.668073] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'boot_index': 0, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281647', 'volume_id': 'ec4ed682-190b-4dbe-b2f8-1b4e3e82fb91', 'name': 'volume-ec4ed682-190b-4dbe-b2f8-1b4e3e82fb91', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '77e19bcd-5963-4c48-bcf2-da5c92b142ff', 'attached_at': '', 'detached_at': '', 'volume_id': 'ec4ed682-190b-4dbe-b2f8-1b4e3e82fb91', 'serial': 'ec4ed682-190b-4dbe-b2f8-1b4e3e82fb91'}, 'device_type': None, 'delete_on_termination': True, 'disk_bus': None, 'mount_device': '/dev/sda', 'guest_format': None, 'attachment_id': '8b042f35-f3bf-454e-b880-851893ecde7c', 'volume_type': None}], 'swap': None} {{(pid=61868) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1290.668293] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Root volume attach. Driver type: vmdk {{(pid=61868) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1290.669039] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0cdfa63-5fb5-4954-b08b-129858abbf13 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.676235] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7f18cfa-4813-4d1a-b1e5-6d722ec8dcce {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.682227] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39e0714b-a5e1-492d-9ca5-ad55a7f6c782 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.688492] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-b60015e2-ad99-4bdd-a12a-85da24eb5bc7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.695040] env[61868]: DEBUG oslo_vmware.api [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1290.695040] env[61868]: value = "task-1316131" [ 1290.695040] env[61868]: _type = "Task" [ 1290.695040] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1290.702678] env[61868]: DEBUG oslo_vmware.api [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316131, 'name': RelocateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1290.835441] env[61868]: DEBUG oslo_vmware.api [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316130, 'name': ReconfigVM_Task, 'duration_secs': 0.294669} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1290.835838] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Reconfigured VM instance instance-0000006c to attach disk [datastore1] e18dfd04-5a37-44d1-930b-52b50362755b/e18dfd04-5a37-44d1-930b-52b50362755b.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1290.836171] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Updating instance 'e18dfd04-5a37-44d1-930b-52b50362755b' progress to 50 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1291.153921] env[61868]: DEBUG nova.network.neutron [req-bd75bc31-7188-4495-a3d6-18e72dc672bd req-b52dd208-d637-4286-8d02-7b4992961a3b service nova] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Updated VIF entry in instance network info cache for port eab3dce4-b36e-4ab7-bc96-27545c10e311. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1291.154345] env[61868]: DEBUG nova.network.neutron [req-bd75bc31-7188-4495-a3d6-18e72dc672bd req-b52dd208-d637-4286-8d02-7b4992961a3b service nova] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Updating instance_info_cache with network_info: [{"id": "eab3dce4-b36e-4ab7-bc96-27545c10e311", "address": "fa:16:3e:4a:7e:64", "network": {"id": "16b3638c-1129-49b8-9f22-636ff7db7386", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1544179589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bdab2203834c4ee5bab11aca70e48cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d298db54-f13d-4bf6-b6c2-755074b3047f", "external-id": "nsx-vlan-transportzone-631", "segmentation_id": 631, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeab3dce4-b3", "ovs_interfaceid": "eab3dce4-b36e-4ab7-bc96-27545c10e311", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1291.209195] env[61868]: DEBUG oslo_vmware.api [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316131, 'name': RelocateVM_Task} progress is 43%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1291.342906] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45d30343-474d-4929-ae56-08bc8f1cbaf1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.366538] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28a03f23-88db-4db1-8d43-1e8e411167e6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.388878] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Updating instance 'e18dfd04-5a37-44d1-930b-52b50362755b' progress to 67 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1291.657697] env[61868]: DEBUG oslo_concurrency.lockutils [req-bd75bc31-7188-4495-a3d6-18e72dc672bd req-b52dd208-d637-4286-8d02-7b4992961a3b service nova] Releasing lock "refresh_cache-77e19bcd-5963-4c48-bcf2-da5c92b142ff" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1291.707713] env[61868]: DEBUG oslo_vmware.api [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316131, 'name': RelocateVM_Task} progress is 58%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1292.208483] env[61868]: DEBUG oslo_vmware.api [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316131, 'name': RelocateVM_Task} progress is 73%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1292.708822] env[61868]: DEBUG oslo_vmware.api [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316131, 'name': RelocateVM_Task} progress is 88%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1293.040212] env[61868]: DEBUG nova.network.neutron [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Port 644d55b3-8c23-4a45-93d6-9d5464d1631e binding to destination host cpu-1 is already ACTIVE {{(pid=61868) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1293.206247] env[61868]: DEBUG oslo_vmware.api [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316131, 'name': RelocateVM_Task} progress is 97%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1293.706615] env[61868]: DEBUG oslo_vmware.api [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316131, 'name': RelocateVM_Task, 'duration_secs': 2.901475} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1293.706910] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Volume attach. Driver type: vmdk {{(pid=61868) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1293.707149] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281647', 'volume_id': 'ec4ed682-190b-4dbe-b2f8-1b4e3e82fb91', 'name': 'volume-ec4ed682-190b-4dbe-b2f8-1b4e3e82fb91', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '77e19bcd-5963-4c48-bcf2-da5c92b142ff', 'attached_at': '', 'detached_at': '', 'volume_id': 'ec4ed682-190b-4dbe-b2f8-1b4e3e82fb91', 'serial': 'ec4ed682-190b-4dbe-b2f8-1b4e3e82fb91'} {{(pid=61868) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1293.707893] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95c040e3-cc28-4edd-9ee0-30020723f58a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.724105] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5648c03-f1ee-4485-ab98-06729da6b4aa {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.744992] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] volume-ec4ed682-190b-4dbe-b2f8-1b4e3e82fb91/volume-ec4ed682-190b-4dbe-b2f8-1b4e3e82fb91.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1293.745206] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e7f5bb68-6030-4d85-abfb-ccaddfb62153 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.762842] env[61868]: DEBUG oslo_vmware.api [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1293.762842] env[61868]: value = "task-1316132" [ 1293.762842] env[61868]: _type = "Task" [ 1293.762842] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1293.769940] env[61868]: DEBUG oslo_vmware.api [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316132, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1294.062485] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "e18dfd04-5a37-44d1-930b-52b50362755b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1294.062742] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "e18dfd04-5a37-44d1-930b-52b50362755b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1294.062936] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "e18dfd04-5a37-44d1-930b-52b50362755b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1294.273936] env[61868]: DEBUG oslo_vmware.api [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316132, 'name': ReconfigVM_Task, 'duration_secs': 0.490476} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1294.274242] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Reconfigured VM instance instance-0000006e to attach disk [datastore2] volume-ec4ed682-190b-4dbe-b2f8-1b4e3e82fb91/volume-ec4ed682-190b-4dbe-b2f8-1b4e3e82fb91.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1294.278875] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-917b374f-13c2-43b5-8682-69ec3da9fcf3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.293137] env[61868]: DEBUG oslo_vmware.api [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1294.293137] env[61868]: value = "task-1316133" [ 1294.293137] env[61868]: _type = "Task" [ 1294.293137] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1294.300662] env[61868]: DEBUG oslo_vmware.api [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316133, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1294.802962] env[61868]: DEBUG oslo_vmware.api [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316133, 'name': ReconfigVM_Task, 'duration_secs': 0.122664} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1294.803336] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281647', 'volume_id': 'ec4ed682-190b-4dbe-b2f8-1b4e3e82fb91', 'name': 'volume-ec4ed682-190b-4dbe-b2f8-1b4e3e82fb91', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '77e19bcd-5963-4c48-bcf2-da5c92b142ff', 'attached_at': '', 'detached_at': '', 'volume_id': 'ec4ed682-190b-4dbe-b2f8-1b4e3e82fb91', 'serial': 'ec4ed682-190b-4dbe-b2f8-1b4e3e82fb91'} {{(pid=61868) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1294.803865] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-09acf0ff-1b22-4fa6-acd0-afa5989c90b1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.811024] env[61868]: DEBUG oslo_vmware.api [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1294.811024] env[61868]: value = "task-1316134" [ 1294.811024] env[61868]: _type = "Task" [ 1294.811024] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1294.819986] env[61868]: DEBUG oslo_vmware.api [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316134, 'name': Rename_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1295.094390] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "refresh_cache-e18dfd04-5a37-44d1-930b-52b50362755b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1295.094587] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquired lock "refresh_cache-e18dfd04-5a37-44d1-930b-52b50362755b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1295.094770] env[61868]: DEBUG nova.network.neutron [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1295.320582] env[61868]: DEBUG oslo_vmware.api [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316134, 'name': Rename_Task, 'duration_secs': 0.127971} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1295.320796] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1295.321046] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9b3ea6f0-f6c6-4c36-8666-59d86e453e27 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.327368] env[61868]: DEBUG oslo_vmware.api [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1295.327368] env[61868]: value = "task-1316135" [ 1295.327368] env[61868]: _type = "Task" [ 1295.327368] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1295.336270] env[61868]: DEBUG oslo_vmware.api [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316135, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1295.814458] env[61868]: DEBUG nova.network.neutron [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Updating instance_info_cache with network_info: [{"id": "644d55b3-8c23-4a45-93d6-9d5464d1631e", "address": "fa:16:3e:db:75:9d", "network": {"id": "7e8cbecf-dfdd-43f2-8745-ae1d2d55d309", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-33935587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a7014141b24101ba7920bc83be4a9e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4db2961d-273d-4634-9d06-a94fa9d384fb", "external-id": "nsx-vlan-transportzone-572", "segmentation_id": 572, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap644d55b3-8c", "ovs_interfaceid": "644d55b3-8c23-4a45-93d6-9d5464d1631e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1295.838118] env[61868]: DEBUG oslo_vmware.api [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316135, 'name': PowerOnVM_Task, 'duration_secs': 0.42712} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1295.838364] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1295.838561] env[61868]: INFO nova.compute.manager [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Took 6.05 seconds to spawn the instance on the hypervisor. [ 1295.838741] env[61868]: DEBUG nova.compute.manager [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1295.839481] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bb07645-865a-49f7-b36a-b0fde75eb627 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1296.317883] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Releasing lock "refresh_cache-e18dfd04-5a37-44d1-930b-52b50362755b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1296.355703] env[61868]: INFO nova.compute.manager [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Took 12.48 seconds to build instance. [ 1296.827611] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86c9c40f-ab74-4f1f-b336-e9753f4732e3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1296.834310] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f42764f0-19c7-46af-b345-be33c56fba69 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1296.857663] env[61868]: DEBUG oslo_concurrency.lockutils [None req-97bd5d42-d82f-40ef-b687-4bdb15515594 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "77e19bcd-5963-4c48-bcf2-da5c92b142ff" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.984s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1297.164591] env[61868]: DEBUG nova.compute.manager [req-b102807b-ab90-4b96-a64f-28bd9aa30254 req-9f0eea11-baa4-4aff-a83b-830c93291a51 service nova] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Received event network-changed-8f9ad20f-f367-4826-8aae-00a1192d0bc8 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1297.164768] env[61868]: DEBUG nova.compute.manager [req-b102807b-ab90-4b96-a64f-28bd9aa30254 req-9f0eea11-baa4-4aff-a83b-830c93291a51 service nova] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Refreshing instance network info cache due to event network-changed-8f9ad20f-f367-4826-8aae-00a1192d0bc8. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1297.164994] env[61868]: DEBUG oslo_concurrency.lockutils [req-b102807b-ab90-4b96-a64f-28bd9aa30254 req-9f0eea11-baa4-4aff-a83b-830c93291a51 service nova] Acquiring lock "refresh_cache-59171bf1-f2d4-4ff5-a661-025f5dca7600" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1297.165143] env[61868]: DEBUG oslo_concurrency.lockutils [req-b102807b-ab90-4b96-a64f-28bd9aa30254 req-9f0eea11-baa4-4aff-a83b-830c93291a51 service nova] Acquired lock "refresh_cache-59171bf1-f2d4-4ff5-a661-025f5dca7600" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1297.165349] env[61868]: DEBUG nova.network.neutron [req-b102807b-ab90-4b96-a64f-28bd9aa30254 req-9f0eea11-baa4-4aff-a83b-830c93291a51 service nova] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Refreshing network info cache for port 8f9ad20f-f367-4826-8aae-00a1192d0bc8 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1297.846793] env[61868]: DEBUG nova.network.neutron [req-b102807b-ab90-4b96-a64f-28bd9aa30254 req-9f0eea11-baa4-4aff-a83b-830c93291a51 service nova] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Updated VIF entry in instance network info cache for port 8f9ad20f-f367-4826-8aae-00a1192d0bc8. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1297.847159] env[61868]: DEBUG nova.network.neutron [req-b102807b-ab90-4b96-a64f-28bd9aa30254 req-9f0eea11-baa4-4aff-a83b-830c93291a51 service nova] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Updating instance_info_cache with network_info: [{"id": "8f9ad20f-f367-4826-8aae-00a1192d0bc8", "address": "fa:16:3e:63:2a:68", "network": {"id": "16b3638c-1129-49b8-9f22-636ff7db7386", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1544179589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bdab2203834c4ee5bab11aca70e48cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d298db54-f13d-4bf6-b6c2-755074b3047f", "external-id": "nsx-vlan-transportzone-631", "segmentation_id": 631, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f9ad20f-f3", "ovs_interfaceid": "8f9ad20f-f367-4826-8aae-00a1192d0bc8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1297.927585] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bd2da18-7b00-48a5-9601-f682f5fe35ee {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.949332] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abe4fd12-cc81-4938-a352-0972798c8fe9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.956476] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Updating instance 'e18dfd04-5a37-44d1-930b-52b50362755b' progress to 83 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1298.350081] env[61868]: DEBUG oslo_concurrency.lockutils [req-b102807b-ab90-4b96-a64f-28bd9aa30254 req-9f0eea11-baa4-4aff-a83b-830c93291a51 service nova] Releasing lock "refresh_cache-59171bf1-f2d4-4ff5-a661-025f5dca7600" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1298.462399] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1298.462736] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e5943466-8c19-4b64-a915-ad25ecb75c0f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.470118] env[61868]: DEBUG oslo_vmware.api [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1298.470118] env[61868]: value = "task-1316136" [ 1298.470118] env[61868]: _type = "Task" [ 1298.470118] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1298.478489] env[61868]: DEBUG oslo_vmware.api [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316136, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1298.529830] env[61868]: DEBUG nova.compute.manager [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Stashing vm_state: active {{(pid=61868) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1298.979936] env[61868]: DEBUG oslo_vmware.api [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316136, 'name': PowerOnVM_Task, 'duration_secs': 0.431667} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1298.980354] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1298.980404] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f53c7401-148f-4919-88e5-a8110a29208a tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Updating instance 'e18dfd04-5a37-44d1-930b-52b50362755b' progress to 100 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1299.047652] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1299.047983] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1299.190259] env[61868]: DEBUG nova.compute.manager [req-52833c37-fdd6-4815-8b4b-47c02569980b req-097d4b40-8b29-4641-b9e4-bb178f14dc0e service nova] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Received event network-changed-eab3dce4-b36e-4ab7-bc96-27545c10e311 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1299.190356] env[61868]: DEBUG nova.compute.manager [req-52833c37-fdd6-4815-8b4b-47c02569980b req-097d4b40-8b29-4641-b9e4-bb178f14dc0e service nova] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Refreshing instance network info cache due to event network-changed-eab3dce4-b36e-4ab7-bc96-27545c10e311. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1299.190575] env[61868]: DEBUG oslo_concurrency.lockutils [req-52833c37-fdd6-4815-8b4b-47c02569980b req-097d4b40-8b29-4641-b9e4-bb178f14dc0e service nova] Acquiring lock "refresh_cache-77e19bcd-5963-4c48-bcf2-da5c92b142ff" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1299.190957] env[61868]: DEBUG oslo_concurrency.lockutils [req-52833c37-fdd6-4815-8b4b-47c02569980b req-097d4b40-8b29-4641-b9e4-bb178f14dc0e service nova] Acquired lock "refresh_cache-77e19bcd-5963-4c48-bcf2-da5c92b142ff" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1299.191229] env[61868]: DEBUG nova.network.neutron [req-52833c37-fdd6-4815-8b4b-47c02569980b req-097d4b40-8b29-4641-b9e4-bb178f14dc0e service nova] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Refreshing network info cache for port eab3dce4-b36e-4ab7-bc96-27545c10e311 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1299.552192] env[61868]: INFO nova.compute.claims [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1299.889271] env[61868]: DEBUG nova.network.neutron [req-52833c37-fdd6-4815-8b4b-47c02569980b req-097d4b40-8b29-4641-b9e4-bb178f14dc0e service nova] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Updated VIF entry in instance network info cache for port eab3dce4-b36e-4ab7-bc96-27545c10e311. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1299.889644] env[61868]: DEBUG nova.network.neutron [req-52833c37-fdd6-4815-8b4b-47c02569980b req-097d4b40-8b29-4641-b9e4-bb178f14dc0e service nova] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Updating instance_info_cache with network_info: [{"id": "eab3dce4-b36e-4ab7-bc96-27545c10e311", "address": "fa:16:3e:4a:7e:64", "network": {"id": "16b3638c-1129-49b8-9f22-636ff7db7386", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1544179589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.190", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bdab2203834c4ee5bab11aca70e48cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d298db54-f13d-4bf6-b6c2-755074b3047f", "external-id": "nsx-vlan-transportzone-631", "segmentation_id": 631, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeab3dce4-b3", "ovs_interfaceid": "eab3dce4-b36e-4ab7-bc96-27545c10e311", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1300.058576] env[61868]: INFO nova.compute.resource_tracker [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Updating resource usage from migration 47069378-6913-48a8-b54c-a6accece8d03 [ 1300.136589] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-817849ac-fb51-4f9d-a26b-51480ae72dc3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.147683] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70e2ce15-d9e5-4796-8c03-79536c82b65f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.190841] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf4413c1-0bcb-4abb-8b09-bff677e54be9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.199543] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-642e7f64-8f91-4ba1-ae73-67dde8f75085 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.213750] env[61868]: DEBUG nova.compute.provider_tree [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1300.392120] env[61868]: DEBUG oslo_concurrency.lockutils [req-52833c37-fdd6-4815-8b4b-47c02569980b req-097d4b40-8b29-4641-b9e4-bb178f14dc0e service nova] Releasing lock "refresh_cache-77e19bcd-5963-4c48-bcf2-da5c92b142ff" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1300.717148] env[61868]: DEBUG nova.scheduler.client.report [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1301.222254] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.174s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1301.222609] env[61868]: INFO nova.compute.manager [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Migrating [ 1301.442604] env[61868]: DEBUG nova.network.neutron [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Port 644d55b3-8c23-4a45-93d6-9d5464d1631e binding to destination host cpu-1 is already ACTIVE {{(pid=61868) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1301.442945] env[61868]: DEBUG oslo_concurrency.lockutils [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "refresh_cache-e18dfd04-5a37-44d1-930b-52b50362755b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1301.443124] env[61868]: DEBUG oslo_concurrency.lockutils [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquired lock "refresh_cache-e18dfd04-5a37-44d1-930b-52b50362755b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1301.443300] env[61868]: DEBUG nova.network.neutron [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1301.742298] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "refresh_cache-77e19bcd-5963-4c48-bcf2-da5c92b142ff" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1301.742577] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquired lock "refresh_cache-77e19bcd-5963-4c48-bcf2-da5c92b142ff" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1301.742677] env[61868]: DEBUG nova.network.neutron [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1302.195497] env[61868]: DEBUG nova.network.neutron [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Updating instance_info_cache with network_info: [{"id": "644d55b3-8c23-4a45-93d6-9d5464d1631e", "address": "fa:16:3e:db:75:9d", "network": {"id": "7e8cbecf-dfdd-43f2-8745-ae1d2d55d309", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-33935587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a7014141b24101ba7920bc83be4a9e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4db2961d-273d-4634-9d06-a94fa9d384fb", "external-id": "nsx-vlan-transportzone-572", "segmentation_id": 572, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap644d55b3-8c", "ovs_interfaceid": "644d55b3-8c23-4a45-93d6-9d5464d1631e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1302.434851] env[61868]: DEBUG nova.network.neutron [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Updating instance_info_cache with network_info: [{"id": "eab3dce4-b36e-4ab7-bc96-27545c10e311", "address": "fa:16:3e:4a:7e:64", "network": {"id": "16b3638c-1129-49b8-9f22-636ff7db7386", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1544179589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.190", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bdab2203834c4ee5bab11aca70e48cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d298db54-f13d-4bf6-b6c2-755074b3047f", "external-id": "nsx-vlan-transportzone-631", "segmentation_id": 631, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeab3dce4-b3", "ovs_interfaceid": "eab3dce4-b36e-4ab7-bc96-27545c10e311", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1302.698568] env[61868]: DEBUG oslo_concurrency.lockutils [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Releasing lock "refresh_cache-e18dfd04-5a37-44d1-930b-52b50362755b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1302.937323] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Releasing lock "refresh_cache-77e19bcd-5963-4c48-bcf2-da5c92b142ff" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1303.202250] env[61868]: DEBUG nova.compute.manager [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=61868) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1304.303389] env[61868]: DEBUG oslo_concurrency.lockutils [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1304.303679] env[61868]: DEBUG oslo_concurrency.lockutils [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1304.450966] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-549906e2-779f-481f-aeb5-28202f7ce5fd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.469857] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Updating instance '77e19bcd-5963-4c48-bcf2-da5c92b142ff' progress to 0 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1304.807134] env[61868]: DEBUG nova.objects.instance [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lazy-loading 'migration_context' on Instance uuid e18dfd04-5a37-44d1-930b-52b50362755b {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1304.975394] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1304.975692] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1777a772-8c4f-4f5c-ac65-8a02bffa0776 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.982741] env[61868]: DEBUG oslo_vmware.api [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1304.982741] env[61868]: value = "task-1316137" [ 1304.982741] env[61868]: _type = "Task" [ 1304.982741] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1304.990783] env[61868]: DEBUG oslo_vmware.api [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316137, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1305.392658] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f6a54f1-8a3d-4443-933e-2834c93cb90e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.400296] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e083cc2-c139-48df-b0ab-0da5360bed16 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.428524] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcc64dc7-a5bc-4bdf-b160-05bd5f6cd3e3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.434886] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98fb5750-570a-4868-9808-5d90f65d0bd5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.447035] env[61868]: DEBUG nova.compute.provider_tree [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1305.490681] env[61868]: DEBUG oslo_vmware.api [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316137, 'name': PowerOffVM_Task, 'duration_secs': 0.207753} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1305.490926] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1305.491124] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Updating instance '77e19bcd-5963-4c48-bcf2-da5c92b142ff' progress to 17 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1305.949925] env[61868]: DEBUG nova.scheduler.client.report [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1305.996532] env[61868]: DEBUG nova.virt.hardware [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1305.996839] env[61868]: DEBUG nova.virt.hardware [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1305.996992] env[61868]: DEBUG nova.virt.hardware [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1305.997196] env[61868]: DEBUG nova.virt.hardware [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1305.997347] env[61868]: DEBUG nova.virt.hardware [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1305.997493] env[61868]: DEBUG nova.virt.hardware [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1305.997706] env[61868]: DEBUG nova.virt.hardware [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1305.997858] env[61868]: DEBUG nova.virt.hardware [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1305.998035] env[61868]: DEBUG nova.virt.hardware [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1305.998203] env[61868]: DEBUG nova.virt.hardware [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1305.998375] env[61868]: DEBUG nova.virt.hardware [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1306.003446] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-be823aa9-b4e5-4a4f-affa-c720415a5618 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.018958] env[61868]: DEBUG oslo_vmware.api [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1306.018958] env[61868]: value = "task-1316138" [ 1306.018958] env[61868]: _type = "Task" [ 1306.018958] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1306.026665] env[61868]: DEBUG oslo_vmware.api [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316138, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1306.529045] env[61868]: DEBUG oslo_vmware.api [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316138, 'name': ReconfigVM_Task, 'duration_secs': 0.16296} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1306.529045] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Updating instance '77e19bcd-5963-4c48-bcf2-da5c92b142ff' progress to 33 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1306.961130] env[61868]: DEBUG oslo_concurrency.lockutils [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.657s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1307.034740] env[61868]: DEBUG nova.virt.hardware [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1307.035032] env[61868]: DEBUG nova.virt.hardware [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1307.035149] env[61868]: DEBUG nova.virt.hardware [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1307.035337] env[61868]: DEBUG nova.virt.hardware [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1307.035488] env[61868]: DEBUG nova.virt.hardware [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1307.035638] env[61868]: DEBUG nova.virt.hardware [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1307.035842] env[61868]: DEBUG nova.virt.hardware [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1307.036015] env[61868]: DEBUG nova.virt.hardware [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1307.036195] env[61868]: DEBUG nova.virt.hardware [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1307.036361] env[61868]: DEBUG nova.virt.hardware [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1307.036545] env[61868]: DEBUG nova.virt.hardware [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1307.041792] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Reconfiguring VM instance instance-0000006e to detach disk 2000 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1307.042393] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9aa3e605-39d7-493e-9aa0-742eaa34df42 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.061697] env[61868]: DEBUG oslo_vmware.api [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1307.061697] env[61868]: value = "task-1316139" [ 1307.061697] env[61868]: _type = "Task" [ 1307.061697] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1307.069463] env[61868]: DEBUG oslo_vmware.api [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316139, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1307.571414] env[61868]: DEBUG oslo_vmware.api [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316139, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1308.073051] env[61868]: DEBUG oslo_vmware.api [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316139, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1308.497892] env[61868]: INFO nova.compute.manager [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Swapping old allocation on dict_keys(['83e67721-2ac9-4a23-aa31-82aca86979c8']) held by migration 0cad9d16-b904-461e-8bbb-0cb210983e12 for instance [ 1308.519365] env[61868]: DEBUG nova.scheduler.client.report [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Overwriting current allocation {'allocations': {'83e67721-2ac9-4a23-aa31-82aca86979c8': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 142}}, 'project_id': '28a7014141b24101ba7920bc83be4a9e', 'user_id': '4a981d6292f34cd7aba2108a9b9f3a90', 'consumer_generation': 1} on consumer e18dfd04-5a37-44d1-930b-52b50362755b {{(pid=61868) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1308.572772] env[61868]: DEBUG oslo_vmware.api [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316139, 'name': ReconfigVM_Task, 'duration_secs': 1.149533} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1308.573063] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Reconfigured VM instance instance-0000006e to detach disk 2000 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1308.573813] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff248b49-d1ca-43d2-90ec-dcdcf306c896 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.596759] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] volume-ec4ed682-190b-4dbe-b2f8-1b4e3e82fb91/volume-ec4ed682-190b-4dbe-b2f8-1b4e3e82fb91.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1308.597731] env[61868]: DEBUG oslo_concurrency.lockutils [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "refresh_cache-e18dfd04-5a37-44d1-930b-52b50362755b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1308.597912] env[61868]: DEBUG oslo_concurrency.lockutils [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquired lock "refresh_cache-e18dfd04-5a37-44d1-930b-52b50362755b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1308.598101] env[61868]: DEBUG nova.network.neutron [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1308.599243] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e7395f72-6920-48eb-8d15-c7d1e6330ecc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.619081] env[61868]: DEBUG oslo_vmware.api [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1308.619081] env[61868]: value = "task-1316140" [ 1308.619081] env[61868]: _type = "Task" [ 1308.619081] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1308.626885] env[61868]: DEBUG oslo_vmware.api [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316140, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1309.128583] env[61868]: DEBUG oslo_vmware.api [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316140, 'name': ReconfigVM_Task, 'duration_secs': 0.236569} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1309.132022] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Reconfigured VM instance instance-0000006e to attach disk [datastore2] volume-ec4ed682-190b-4dbe-b2f8-1b4e3e82fb91/volume-ec4ed682-190b-4dbe-b2f8-1b4e3e82fb91.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1309.132022] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Updating instance '77e19bcd-5963-4c48-bcf2-da5c92b142ff' progress to 50 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1309.313063] env[61868]: DEBUG nova.network.neutron [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Updating instance_info_cache with network_info: [{"id": "644d55b3-8c23-4a45-93d6-9d5464d1631e", "address": "fa:16:3e:db:75:9d", "network": {"id": "7e8cbecf-dfdd-43f2-8745-ae1d2d55d309", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-33935587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a7014141b24101ba7920bc83be4a9e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4db2961d-273d-4634-9d06-a94fa9d384fb", "external-id": "nsx-vlan-transportzone-572", "segmentation_id": 572, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap644d55b3-8c", "ovs_interfaceid": "644d55b3-8c23-4a45-93d6-9d5464d1631e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1309.636553] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9451b935-53a8-4351-a79d-7062e49a64c1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.655463] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0a4e60d-e42f-463d-9722-36f1a452b319 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.672007] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Updating instance '77e19bcd-5963-4c48-bcf2-da5c92b142ff' progress to 67 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1309.815262] env[61868]: DEBUG oslo_concurrency.lockutils [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Releasing lock "refresh_cache-e18dfd04-5a37-44d1-930b-52b50362755b" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1309.816271] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83dc60f6-6581-494c-a5c6-91a238c0f129 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.823391] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8850fda6-ecc7-4136-b68e-db026a2b19c4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.902605] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1310.903038] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-495e8997-bc83-4575-b718-839ea0bf67ab {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.910706] env[61868]: DEBUG oslo_vmware.api [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1310.910706] env[61868]: value = "task-1316141" [ 1310.910706] env[61868]: _type = "Task" [ 1310.910706] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1310.918162] env[61868]: DEBUG oslo_vmware.api [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316141, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1311.302980] env[61868]: DEBUG nova.network.neutron [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Port eab3dce4-b36e-4ab7-bc96-27545c10e311 binding to destination host cpu-1 is already ACTIVE {{(pid=61868) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1311.420119] env[61868]: DEBUG oslo_vmware.api [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316141, 'name': PowerOffVM_Task, 'duration_secs': 0.199992} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1311.420394] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1311.421044] env[61868]: DEBUG nova.virt.hardware [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1311.421266] env[61868]: DEBUG nova.virt.hardware [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1311.421426] env[61868]: DEBUG nova.virt.hardware [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1311.421612] env[61868]: DEBUG nova.virt.hardware [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1311.421762] env[61868]: DEBUG nova.virt.hardware [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1311.421911] env[61868]: DEBUG nova.virt.hardware [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1311.422144] env[61868]: DEBUG nova.virt.hardware [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1311.422316] env[61868]: DEBUG nova.virt.hardware [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1311.422487] env[61868]: DEBUG nova.virt.hardware [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1311.422653] env[61868]: DEBUG nova.virt.hardware [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1311.422842] env[61868]: DEBUG nova.virt.hardware [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1311.427754] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d563f829-d35a-4da6-b1bf-1100d10bc42a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.442260] env[61868]: DEBUG oslo_vmware.api [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1311.442260] env[61868]: value = "task-1316142" [ 1311.442260] env[61868]: _type = "Task" [ 1311.442260] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1311.449458] env[61868]: DEBUG oslo_vmware.api [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316142, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1311.952791] env[61868]: DEBUG oslo_vmware.api [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316142, 'name': ReconfigVM_Task, 'duration_secs': 0.132533} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1311.953595] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15993e74-56cc-4379-ac66-6438ecf7a841 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.973730] env[61868]: DEBUG nova.virt.hardware [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1311.973968] env[61868]: DEBUG nova.virt.hardware [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1311.974149] env[61868]: DEBUG nova.virt.hardware [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1311.974338] env[61868]: DEBUG nova.virt.hardware [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1311.974491] env[61868]: DEBUG nova.virt.hardware [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1311.974644] env[61868]: DEBUG nova.virt.hardware [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1311.974872] env[61868]: DEBUG nova.virt.hardware [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1311.975053] env[61868]: DEBUG nova.virt.hardware [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1311.975230] env[61868]: DEBUG nova.virt.hardware [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1311.975396] env[61868]: DEBUG nova.virt.hardware [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1311.975574] env[61868]: DEBUG nova.virt.hardware [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1311.976349] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b220f83-87b9-4daa-a84d-f21b634e4895 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.982244] env[61868]: DEBUG oslo_vmware.api [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1311.982244] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5214091c-401d-adca-abb6-cd9e552395e9" [ 1311.982244] env[61868]: _type = "Task" [ 1311.982244] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1311.989717] env[61868]: DEBUG oslo_vmware.api [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5214091c-401d-adca-abb6-cd9e552395e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1312.324381] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "77e19bcd-5963-4c48-bcf2-da5c92b142ff-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1312.324638] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "77e19bcd-5963-4c48-bcf2-da5c92b142ff-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1312.324726] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "77e19bcd-5963-4c48-bcf2-da5c92b142ff-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1312.493687] env[61868]: DEBUG oslo_vmware.api [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5214091c-401d-adca-abb6-cd9e552395e9, 'name': SearchDatastore_Task, 'duration_secs': 0.009192} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1312.498896] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Reconfiguring VM instance instance-0000006c to detach disk 2000 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1312.499212] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d5613af1-c98b-4340-98a9-597a3c4e930e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.516371] env[61868]: DEBUG oslo_vmware.api [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1312.516371] env[61868]: value = "task-1316143" [ 1312.516371] env[61868]: _type = "Task" [ 1312.516371] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1312.526157] env[61868]: DEBUG oslo_vmware.api [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316143, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1313.027062] env[61868]: DEBUG oslo_vmware.api [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316143, 'name': ReconfigVM_Task, 'duration_secs': 0.231005} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1313.027062] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Reconfigured VM instance instance-0000006c to detach disk 2000 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1313.027526] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d767407-8d54-4d5e-92e4-484af2973265 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.051015] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] e18dfd04-5a37-44d1-930b-52b50362755b/e18dfd04-5a37-44d1-930b-52b50362755b.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1313.051306] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c1ec9077-435f-4b10-b10c-f9b067ecbf9f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.068449] env[61868]: DEBUG oslo_vmware.api [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1313.068449] env[61868]: value = "task-1316144" [ 1313.068449] env[61868]: _type = "Task" [ 1313.068449] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1313.077022] env[61868]: DEBUG oslo_vmware.api [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316144, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1313.359114] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "refresh_cache-77e19bcd-5963-4c48-bcf2-da5c92b142ff" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1313.359324] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquired lock "refresh_cache-77e19bcd-5963-4c48-bcf2-da5c92b142ff" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1313.359508] env[61868]: DEBUG nova.network.neutron [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1313.578016] env[61868]: DEBUG oslo_vmware.api [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316144, 'name': ReconfigVM_Task, 'duration_secs': 0.262754} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1313.578327] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Reconfigured VM instance instance-0000006c to attach disk [datastore1] e18dfd04-5a37-44d1-930b-52b50362755b/e18dfd04-5a37-44d1-930b-52b50362755b.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1313.579140] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-991253a3-edd3-45b4-98ee-a0c33677d096 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.598365] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7bcc844-897e-485a-8016-454b273db900 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.619237] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66e1a5b4-9ff3-425b-a4ca-bfcfd75207f1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.639786] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a29a3f25-2010-46dc-a92d-80907ce6f30e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.645863] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1313.646083] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9a7134f8-ddf6-4967-bfec-e6d02dedbaa7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.651692] env[61868]: DEBUG oslo_vmware.api [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1313.651692] env[61868]: value = "task-1316145" [ 1313.651692] env[61868]: _type = "Task" [ 1313.651692] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1313.658758] env[61868]: DEBUG oslo_vmware.api [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316145, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1314.113059] env[61868]: DEBUG nova.network.neutron [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Updating instance_info_cache with network_info: [{"id": "eab3dce4-b36e-4ab7-bc96-27545c10e311", "address": "fa:16:3e:4a:7e:64", "network": {"id": "16b3638c-1129-49b8-9f22-636ff7db7386", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1544179589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.190", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bdab2203834c4ee5bab11aca70e48cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d298db54-f13d-4bf6-b6c2-755074b3047f", "external-id": "nsx-vlan-transportzone-631", "segmentation_id": 631, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeab3dce4-b3", "ovs_interfaceid": "eab3dce4-b36e-4ab7-bc96-27545c10e311", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1314.161429] env[61868]: DEBUG oslo_vmware.api [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316145, 'name': PowerOnVM_Task, 'duration_secs': 0.348052} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1314.162996] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1314.616341] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Releasing lock "refresh_cache-77e19bcd-5963-4c48-bcf2-da5c92b142ff" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1315.127037] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f4f25e9-e042-4428-85c5-46b17de890b9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.133270] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f178cf1a-800a-4e90-8273-3f1518321b28 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.203633] env[61868]: INFO nova.compute.manager [None req-631ceb22-b5c4-4baf-86bf-f18fff4f092f tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Updating instance to original state: 'active' [ 1316.233406] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db5d3b2d-3092-4666-b0fd-6a8cf0f54f8d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.254640] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-432e4833-c104-4cab-9b66-1e12fca75719 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.261919] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Updating instance '77e19bcd-5963-4c48-bcf2-da5c92b142ff' progress to 83 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1316.769033] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1316.769033] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d48b0654-0a8e-44f7-bfd3-a7b7799a2765 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.776522] env[61868]: DEBUG oslo_vmware.api [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1316.776522] env[61868]: value = "task-1316146" [ 1316.776522] env[61868]: _type = "Task" [ 1316.776522] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1316.784998] env[61868]: DEBUG oslo_vmware.api [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316146, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1317.152531] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "e18dfd04-5a37-44d1-930b-52b50362755b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1317.152842] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "e18dfd04-5a37-44d1-930b-52b50362755b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1317.153107] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "e18dfd04-5a37-44d1-930b-52b50362755b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1317.153310] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "e18dfd04-5a37-44d1-930b-52b50362755b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1317.153488] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "e18dfd04-5a37-44d1-930b-52b50362755b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1317.155814] env[61868]: INFO nova.compute.manager [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Terminating instance [ 1317.157920] env[61868]: DEBUG nova.compute.manager [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1317.158155] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1317.158392] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ad441b62-d8d1-4490-a4ac-40a65691e98d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1317.167031] env[61868]: DEBUG oslo_vmware.api [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1317.167031] env[61868]: value = "task-1316147" [ 1317.167031] env[61868]: _type = "Task" [ 1317.167031] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1317.176731] env[61868]: DEBUG oslo_vmware.api [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316147, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1317.286584] env[61868]: DEBUG oslo_vmware.api [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316146, 'name': PowerOnVM_Task, 'duration_secs': 0.350412} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1317.287051] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1317.287051] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-bb5ac060-35db-4136-a1f1-d73d49201c26 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Updating instance '77e19bcd-5963-4c48-bcf2-da5c92b142ff' progress to 100 {{(pid=61868) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1317.676129] env[61868]: DEBUG oslo_vmware.api [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316147, 'name': PowerOffVM_Task, 'duration_secs': 0.256421} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1317.676384] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1317.676583] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Volume detach. Driver type: vmdk {{(pid=61868) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1317.676811] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281646', 'volume_id': '6bb58fe7-5636-40f7-8375-5dd71b90fca3', 'name': 'volume-6bb58fe7-5636-40f7-8375-5dd71b90fca3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'e18dfd04-5a37-44d1-930b-52b50362755b', 'attached_at': '2024-10-18T17:31:58.000000', 'detached_at': '', 'volume_id': '6bb58fe7-5636-40f7-8375-5dd71b90fca3', 'serial': '6bb58fe7-5636-40f7-8375-5dd71b90fca3'} {{(pid=61868) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1317.677553] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2def7d0-7fae-4b61-bc0e-b34567e2a18c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1317.697717] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4e307b5-2e14-4681-b842-ae97422c1545 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1317.703436] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74623b9f-20a3-4b7f-be15-e8c0c1c8d21b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1317.723401] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5929de83-65dd-4071-8075-beed288603ea {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1317.737136] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] The volume has not been displaced from its original location: [datastore1] volume-6bb58fe7-5636-40f7-8375-5dd71b90fca3/volume-6bb58fe7-5636-40f7-8375-5dd71b90fca3.vmdk. No consolidation needed. {{(pid=61868) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1317.742175] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Reconfiguring VM instance instance-0000006c to detach disk 2001 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1317.742418] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-75375bb9-020b-4095-864d-dba0489fb3eb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1317.758981] env[61868]: DEBUG oslo_vmware.api [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1317.758981] env[61868]: value = "task-1316148" [ 1317.758981] env[61868]: _type = "Task" [ 1317.758981] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1317.767996] env[61868]: DEBUG oslo_vmware.api [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316148, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1318.272396] env[61868]: DEBUG oslo_vmware.api [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316148, 'name': ReconfigVM_Task, 'duration_secs': 0.194721} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1318.272791] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Reconfigured VM instance instance-0000006c to detach disk 2001 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1318.280169] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6dfca8b7-2249-46e9-a3bd-1999fde12313 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.301249] env[61868]: DEBUG oslo_vmware.api [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1318.301249] env[61868]: value = "task-1316149" [ 1318.301249] env[61868]: _type = "Task" [ 1318.301249] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1318.311742] env[61868]: DEBUG oslo_vmware.api [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316149, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1318.811096] env[61868]: DEBUG oslo_vmware.api [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316149, 'name': ReconfigVM_Task, 'duration_secs': 0.133432} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1318.811355] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281646', 'volume_id': '6bb58fe7-5636-40f7-8375-5dd71b90fca3', 'name': 'volume-6bb58fe7-5636-40f7-8375-5dd71b90fca3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'e18dfd04-5a37-44d1-930b-52b50362755b', 'attached_at': '2024-10-18T17:31:58.000000', 'detached_at': '', 'volume_id': '6bb58fe7-5636-40f7-8375-5dd71b90fca3', 'serial': '6bb58fe7-5636-40f7-8375-5dd71b90fca3'} {{(pid=61868) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1318.811644] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1318.812385] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9792ba71-41d0-4574-a8c1-54b4b3891497 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.818771] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1318.818977] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-af83e6c0-ac54-42b4-a1a9-f62268a68e36 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.879676] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1318.879909] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1318.880109] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Deleting the datastore file [datastore1] e18dfd04-5a37-44d1-930b-52b50362755b {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1318.880371] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9ba11a59-ece3-4262-b2f1-600847846b74 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.886838] env[61868]: DEBUG oslo_vmware.api [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1318.886838] env[61868]: value = "task-1316151" [ 1318.886838] env[61868]: _type = "Task" [ 1318.886838] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1318.894206] env[61868]: DEBUG oslo_vmware.api [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316151, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1319.076055] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d6ec80f8-827e-49b2-a427-3c1f0dc84d7f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "77e19bcd-5963-4c48-bcf2-da5c92b142ff" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1319.076306] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d6ec80f8-827e-49b2-a427-3c1f0dc84d7f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "77e19bcd-5963-4c48-bcf2-da5c92b142ff" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1319.076619] env[61868]: DEBUG nova.compute.manager [None req-d6ec80f8-827e-49b2-a427-3c1f0dc84d7f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Going to confirm migration 6 {{(pid=61868) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1319.399630] env[61868]: DEBUG oslo_vmware.api [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316151, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141206} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1319.400034] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1319.400245] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1319.400543] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1319.400839] env[61868]: INFO nova.compute.manager [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Took 2.24 seconds to destroy the instance on the hypervisor. [ 1319.401205] env[61868]: DEBUG oslo.service.loopingcall [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1319.401430] env[61868]: DEBUG nova.compute.manager [-] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1319.401528] env[61868]: DEBUG nova.network.neutron [-] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1319.642957] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d6ec80f8-827e-49b2-a427-3c1f0dc84d7f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "refresh_cache-77e19bcd-5963-4c48-bcf2-da5c92b142ff" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1319.643178] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d6ec80f8-827e-49b2-a427-3c1f0dc84d7f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquired lock "refresh_cache-77e19bcd-5963-4c48-bcf2-da5c92b142ff" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1319.643367] env[61868]: DEBUG nova.network.neutron [None req-d6ec80f8-827e-49b2-a427-3c1f0dc84d7f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1319.643553] env[61868]: DEBUG nova.objects.instance [None req-d6ec80f8-827e-49b2-a427-3c1f0dc84d7f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lazy-loading 'info_cache' on Instance uuid 77e19bcd-5963-4c48-bcf2-da5c92b142ff {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1319.867900] env[61868]: DEBUG nova.compute.manager [req-062b7a6e-e7d6-4928-bfa3-55b8aa667a41 req-0671328f-1c66-4980-ad0f-4d95e6bf3fb5 service nova] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Received event network-vif-deleted-644d55b3-8c23-4a45-93d6-9d5464d1631e {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1319.868169] env[61868]: INFO nova.compute.manager [req-062b7a6e-e7d6-4928-bfa3-55b8aa667a41 req-0671328f-1c66-4980-ad0f-4d95e6bf3fb5 service nova] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Neutron deleted interface 644d55b3-8c23-4a45-93d6-9d5464d1631e; detaching it from the instance and deleting it from the info cache [ 1319.868312] env[61868]: DEBUG nova.network.neutron [req-062b7a6e-e7d6-4928-bfa3-55b8aa667a41 req-0671328f-1c66-4980-ad0f-4d95e6bf3fb5 service nova] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1320.348244] env[61868]: DEBUG nova.network.neutron [-] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1320.371626] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a4a3d457-0bf4-4822-987e-bfad14537270 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.381925] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f14756de-5223-4344-ae74-30ba06418e61 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.406335] env[61868]: DEBUG nova.compute.manager [req-062b7a6e-e7d6-4928-bfa3-55b8aa667a41 req-0671328f-1c66-4980-ad0f-4d95e6bf3fb5 service nova] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Detach interface failed, port_id=644d55b3-8c23-4a45-93d6-9d5464d1631e, reason: Instance e18dfd04-5a37-44d1-930b-52b50362755b could not be found. {{(pid=61868) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1320.850998] env[61868]: INFO nova.compute.manager [-] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Took 1.45 seconds to deallocate network for instance. [ 1320.897391] env[61868]: DEBUG nova.network.neutron [None req-d6ec80f8-827e-49b2-a427-3c1f0dc84d7f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Updating instance_info_cache with network_info: [{"id": "eab3dce4-b36e-4ab7-bc96-27545c10e311", "address": "fa:16:3e:4a:7e:64", "network": {"id": "16b3638c-1129-49b8-9f22-636ff7db7386", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1544179589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.190", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bdab2203834c4ee5bab11aca70e48cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d298db54-f13d-4bf6-b6c2-755074b3047f", "external-id": "nsx-vlan-transportzone-631", "segmentation_id": 631, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeab3dce4-b3", "ovs_interfaceid": "eab3dce4-b36e-4ab7-bc96-27545c10e311", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1321.392909] env[61868]: INFO nova.compute.manager [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Took 0.54 seconds to detach 1 volumes for instance. [ 1321.400122] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d6ec80f8-827e-49b2-a427-3c1f0dc84d7f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Releasing lock "refresh_cache-77e19bcd-5963-4c48-bcf2-da5c92b142ff" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1321.400362] env[61868]: DEBUG nova.objects.instance [None req-d6ec80f8-827e-49b2-a427-3c1f0dc84d7f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lazy-loading 'migration_context' on Instance uuid 77e19bcd-5963-4c48-bcf2-da5c92b142ff {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1321.899867] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1321.900361] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1321.900361] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1321.903008] env[61868]: DEBUG nova.objects.base [None req-d6ec80f8-827e-49b2-a427-3c1f0dc84d7f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Object Instance<77e19bcd-5963-4c48-bcf2-da5c92b142ff> lazy-loaded attributes: info_cache,migration_context {{(pid=61868) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1321.903878] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7a3975f-dde4-4c0c-8486-506a860028fc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.925350] env[61868]: INFO nova.scheduler.client.report [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Deleted allocations for instance e18dfd04-5a37-44d1-930b-52b50362755b [ 1321.926399] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-84366bc0-232b-4eda-9489-2b63e598db6e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.934150] env[61868]: DEBUG oslo_vmware.api [None req-d6ec80f8-827e-49b2-a427-3c1f0dc84d7f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1321.934150] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52445744-d0e8-3a39-e24e-f08eb859ad2c" [ 1321.934150] env[61868]: _type = "Task" [ 1321.934150] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1321.948654] env[61868]: DEBUG oslo_vmware.api [None req-d6ec80f8-827e-49b2-a427-3c1f0dc84d7f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52445744-d0e8-3a39-e24e-f08eb859ad2c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1322.434754] env[61868]: DEBUG oslo_concurrency.lockutils [None req-f4cab554-052a-4505-a58d-93e0503e7c52 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "e18dfd04-5a37-44d1-930b-52b50362755b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.282s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1322.444729] env[61868]: DEBUG oslo_vmware.api [None req-d6ec80f8-827e-49b2-a427-3c1f0dc84d7f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52445744-d0e8-3a39-e24e-f08eb859ad2c, 'name': SearchDatastore_Task, 'duration_secs': 0.018319} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1322.444985] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d6ec80f8-827e-49b2-a427-3c1f0dc84d7f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1322.445305] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d6ec80f8-827e-49b2-a427-3c1f0dc84d7f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1323.009010] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f26500d-3fa2-4fa2-92c0-2f2af3a6af48 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.017298] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d337d7f0-8649-4d05-ad3b-17385e55c64d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.046430] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-781818e5-0ac4-48f0-b4e4-54a76ac46d71 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.054138] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e963da5-4e6d-4a58-93f7-99ee6cfbfef5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.068545] env[61868]: DEBUG nova.compute.provider_tree [None req-d6ec80f8-827e-49b2-a427-3c1f0dc84d7f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1323.112605] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "b411db59-3058-4aa8-ac33-d3897c36c123" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1323.112894] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "b411db59-3058-4aa8-ac33-d3897c36c123" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1323.571855] env[61868]: DEBUG nova.scheduler.client.report [None req-d6ec80f8-827e-49b2-a427-3c1f0dc84d7f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1323.614799] env[61868]: DEBUG nova.compute.manager [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1324.140218] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1324.581831] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d6ec80f8-827e-49b2-a427-3c1f0dc84d7f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.136s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1324.584613] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.445s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1324.586090] env[61868]: INFO nova.compute.claims [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1325.139039] env[61868]: INFO nova.scheduler.client.report [None req-d6ec80f8-827e-49b2-a427-3c1f0dc84d7f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Deleted allocation for migration 47069378-6913-48a8-b54c-a6accece8d03 [ 1325.644737] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d6ec80f8-827e-49b2-a427-3c1f0dc84d7f tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "77e19bcd-5963-4c48-bcf2-da5c92b142ff" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.568s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1325.658466] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b652fa71-0be7-408d-b900-239e480949fb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.667110] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21b45982-cce2-4010-8652-bd44f839e506 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.674676] env[61868]: INFO nova.compute.manager [None req-7f60677b-20b6-44d5-8439-a58a337f80d0 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Get console output [ 1325.675045] env[61868]: WARNING nova.virt.vmwareapi.driver [None req-7f60677b-20b6-44d5-8439-a58a337f80d0 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] The console log is missing. Check your VSPC configuration [ 1325.701763] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc7378fd-cc73-40e6-902d-04e97619140a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.709758] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48535151-5eb0-48d9-9a59-37296689c082 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.723187] env[61868]: DEBUG nova.compute.provider_tree [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1326.226389] env[61868]: DEBUG nova.scheduler.client.report [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1326.731744] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.147s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1326.732305] env[61868]: DEBUG nova.compute.manager [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1327.237228] env[61868]: DEBUG nova.compute.utils [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1327.238674] env[61868]: DEBUG nova.compute.manager [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1327.238844] env[61868]: DEBUG nova.network.neutron [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1327.286420] env[61868]: DEBUG nova.policy [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4a981d6292f34cd7aba2108a9b9f3a90', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '28a7014141b24101ba7920bc83be4a9e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 1327.540714] env[61868]: DEBUG nova.network.neutron [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Successfully created port: 22435bb8-30a0-4fbb-b86c-a27082b82e4a {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1327.742783] env[61868]: DEBUG nova.compute.manager [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1328.752990] env[61868]: DEBUG nova.compute.manager [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1328.777926] env[61868]: DEBUG nova.virt.hardware [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1328.778223] env[61868]: DEBUG nova.virt.hardware [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1328.778400] env[61868]: DEBUG nova.virt.hardware [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1328.778587] env[61868]: DEBUG nova.virt.hardware [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1328.778736] env[61868]: DEBUG nova.virt.hardware [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1328.778884] env[61868]: DEBUG nova.virt.hardware [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1328.779109] env[61868]: DEBUG nova.virt.hardware [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1328.779272] env[61868]: DEBUG nova.virt.hardware [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1328.779441] env[61868]: DEBUG nova.virt.hardware [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1328.779606] env[61868]: DEBUG nova.virt.hardware [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1328.779781] env[61868]: DEBUG nova.virt.hardware [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1328.780682] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7b4c7ba-c3fd-4d2d-81aa-60e6c981cc50 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.789560] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-750b14cd-d10b-4b8e-9e0c-07798c5887ab {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.911767] env[61868]: DEBUG nova.compute.manager [req-159c3917-cd10-4d50-b7b1-0524f97f2ac2 req-d551940f-79dc-495e-96f9-777d8bfecb05 service nova] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Received event network-vif-plugged-22435bb8-30a0-4fbb-b86c-a27082b82e4a {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1328.912013] env[61868]: DEBUG oslo_concurrency.lockutils [req-159c3917-cd10-4d50-b7b1-0524f97f2ac2 req-d551940f-79dc-495e-96f9-777d8bfecb05 service nova] Acquiring lock "b411db59-3058-4aa8-ac33-d3897c36c123-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1328.912332] env[61868]: DEBUG oslo_concurrency.lockutils [req-159c3917-cd10-4d50-b7b1-0524f97f2ac2 req-d551940f-79dc-495e-96f9-777d8bfecb05 service nova] Lock "b411db59-3058-4aa8-ac33-d3897c36c123-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1328.912422] env[61868]: DEBUG oslo_concurrency.lockutils [req-159c3917-cd10-4d50-b7b1-0524f97f2ac2 req-d551940f-79dc-495e-96f9-777d8bfecb05 service nova] Lock "b411db59-3058-4aa8-ac33-d3897c36c123-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1328.912559] env[61868]: DEBUG nova.compute.manager [req-159c3917-cd10-4d50-b7b1-0524f97f2ac2 req-d551940f-79dc-495e-96f9-777d8bfecb05 service nova] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] No waiting events found dispatching network-vif-plugged-22435bb8-30a0-4fbb-b86c-a27082b82e4a {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1328.912718] env[61868]: WARNING nova.compute.manager [req-159c3917-cd10-4d50-b7b1-0524f97f2ac2 req-d551940f-79dc-495e-96f9-777d8bfecb05 service nova] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Received unexpected event network-vif-plugged-22435bb8-30a0-4fbb-b86c-a27082b82e4a for instance with vm_state building and task_state spawning. [ 1329.044866] env[61868]: DEBUG nova.network.neutron [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Successfully updated port: 22435bb8-30a0-4fbb-b86c-a27082b82e4a {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1329.547648] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "refresh_cache-b411db59-3058-4aa8-ac33-d3897c36c123" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1329.547848] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquired lock "refresh_cache-b411db59-3058-4aa8-ac33-d3897c36c123" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1329.547934] env[61868]: DEBUG nova.network.neutron [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1330.095744] env[61868]: DEBUG nova.network.neutron [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1330.237385] env[61868]: DEBUG nova.network.neutron [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Updating instance_info_cache with network_info: [{"id": "22435bb8-30a0-4fbb-b86c-a27082b82e4a", "address": "fa:16:3e:89:f5:a6", "network": {"id": "7e8cbecf-dfdd-43f2-8745-ae1d2d55d309", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-33935587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a7014141b24101ba7920bc83be4a9e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4db2961d-273d-4634-9d06-a94fa9d384fb", "external-id": "nsx-vlan-transportzone-572", "segmentation_id": 572, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22435bb8-30", "ovs_interfaceid": "22435bb8-30a0-4fbb-b86c-a27082b82e4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1330.740297] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Releasing lock "refresh_cache-b411db59-3058-4aa8-ac33-d3897c36c123" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1330.740623] env[61868]: DEBUG nova.compute.manager [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Instance network_info: |[{"id": "22435bb8-30a0-4fbb-b86c-a27082b82e4a", "address": "fa:16:3e:89:f5:a6", "network": {"id": "7e8cbecf-dfdd-43f2-8745-ae1d2d55d309", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-33935587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a7014141b24101ba7920bc83be4a9e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4db2961d-273d-4634-9d06-a94fa9d384fb", "external-id": "nsx-vlan-transportzone-572", "segmentation_id": 572, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22435bb8-30", "ovs_interfaceid": "22435bb8-30a0-4fbb-b86c-a27082b82e4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1330.741086] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:89:f5:a6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4db2961d-273d-4634-9d06-a94fa9d384fb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '22435bb8-30a0-4fbb-b86c-a27082b82e4a', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1330.748593] env[61868]: DEBUG oslo.service.loopingcall [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1330.748795] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1330.749044] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-022fde97-7afe-472c-bf44-edecc0f9b250 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.768934] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1330.768934] env[61868]: value = "task-1316152" [ 1330.768934] env[61868]: _type = "Task" [ 1330.768934] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1330.776254] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1316152, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1330.938416] env[61868]: DEBUG nova.compute.manager [req-05e57def-efa0-4001-9c35-8368092d384c req-e789da67-1daa-45ea-9745-13560a14315c service nova] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Received event network-changed-22435bb8-30a0-4fbb-b86c-a27082b82e4a {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1330.938703] env[61868]: DEBUG nova.compute.manager [req-05e57def-efa0-4001-9c35-8368092d384c req-e789da67-1daa-45ea-9745-13560a14315c service nova] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Refreshing instance network info cache due to event network-changed-22435bb8-30a0-4fbb-b86c-a27082b82e4a. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1330.938931] env[61868]: DEBUG oslo_concurrency.lockutils [req-05e57def-efa0-4001-9c35-8368092d384c req-e789da67-1daa-45ea-9745-13560a14315c service nova] Acquiring lock "refresh_cache-b411db59-3058-4aa8-ac33-d3897c36c123" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1330.939096] env[61868]: DEBUG oslo_concurrency.lockutils [req-05e57def-efa0-4001-9c35-8368092d384c req-e789da67-1daa-45ea-9745-13560a14315c service nova] Acquired lock "refresh_cache-b411db59-3058-4aa8-ac33-d3897c36c123" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1330.939285] env[61868]: DEBUG nova.network.neutron [req-05e57def-efa0-4001-9c35-8368092d384c req-e789da67-1daa-45ea-9745-13560a14315c service nova] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Refreshing network info cache for port 22435bb8-30a0-4fbb-b86c-a27082b82e4a {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1331.279320] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1316152, 'name': CreateVM_Task, 'duration_secs': 0.270709} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1331.279630] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1331.280170] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1331.280353] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1331.280669] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1331.280914] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38c230bf-def7-4554-847d-e15e2b813345 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.285077] env[61868]: DEBUG oslo_vmware.api [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1331.285077] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5250eff2-eac7-079f-4bee-0e09bad4fc49" [ 1331.285077] env[61868]: _type = "Task" [ 1331.285077] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1331.292185] env[61868]: DEBUG oslo_vmware.api [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5250eff2-eac7-079f-4bee-0e09bad4fc49, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1331.615437] env[61868]: DEBUG nova.network.neutron [req-05e57def-efa0-4001-9c35-8368092d384c req-e789da67-1daa-45ea-9745-13560a14315c service nova] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Updated VIF entry in instance network info cache for port 22435bb8-30a0-4fbb-b86c-a27082b82e4a. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1331.615785] env[61868]: DEBUG nova.network.neutron [req-05e57def-efa0-4001-9c35-8368092d384c req-e789da67-1daa-45ea-9745-13560a14315c service nova] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Updating instance_info_cache with network_info: [{"id": "22435bb8-30a0-4fbb-b86c-a27082b82e4a", "address": "fa:16:3e:89:f5:a6", "network": {"id": "7e8cbecf-dfdd-43f2-8745-ae1d2d55d309", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-33935587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a7014141b24101ba7920bc83be4a9e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4db2961d-273d-4634-9d06-a94fa9d384fb", "external-id": "nsx-vlan-transportzone-572", "segmentation_id": 572, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22435bb8-30", "ovs_interfaceid": "22435bb8-30a0-4fbb-b86c-a27082b82e4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1331.795791] env[61868]: DEBUG oslo_vmware.api [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5250eff2-eac7-079f-4bee-0e09bad4fc49, 'name': SearchDatastore_Task, 'duration_secs': 0.010378} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1331.796045] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1331.796286] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1331.796517] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1331.796665] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1331.796839] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1331.797133] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d53549e1-b5ca-4a62-96db-7bfd9b6c2530 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.806177] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1331.806355] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1331.807037] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b9f48224-137f-4ec7-a773-7bbcc2e53072 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.811497] env[61868]: DEBUG oslo_vmware.api [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1331.811497] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]521f9d15-961b-352a-adf3-7a4d5421f0ed" [ 1331.811497] env[61868]: _type = "Task" [ 1331.811497] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1331.818500] env[61868]: DEBUG oslo_vmware.api [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]521f9d15-961b-352a-adf3-7a4d5421f0ed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1332.118634] env[61868]: DEBUG oslo_concurrency.lockutils [req-05e57def-efa0-4001-9c35-8368092d384c req-e789da67-1daa-45ea-9745-13560a14315c service nova] Releasing lock "refresh_cache-b411db59-3058-4aa8-ac33-d3897c36c123" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1332.322151] env[61868]: DEBUG oslo_vmware.api [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]521f9d15-961b-352a-adf3-7a4d5421f0ed, 'name': SearchDatastore_Task, 'duration_secs': 0.009885} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1332.322809] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c609f3d2-26a0-465b-9ab6-2d4f40ad655d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.327556] env[61868]: DEBUG oslo_vmware.api [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1332.327556] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]529bfe68-9a87-702a-5470-676d7645c260" [ 1332.327556] env[61868]: _type = "Task" [ 1332.327556] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1332.334739] env[61868]: DEBUG oslo_vmware.api [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]529bfe68-9a87-702a-5470-676d7645c260, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1332.838207] env[61868]: DEBUG oslo_vmware.api [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]529bfe68-9a87-702a-5470-676d7645c260, 'name': SearchDatastore_Task, 'duration_secs': 0.00909} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1332.838449] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1332.838703] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] b411db59-3058-4aa8-ac33-d3897c36c123/b411db59-3058-4aa8-ac33-d3897c36c123.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1332.838953] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-508e9024-0f7a-4f3c-9acf-86a79cb1cb99 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.844841] env[61868]: DEBUG oslo_vmware.api [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1332.844841] env[61868]: value = "task-1316153" [ 1332.844841] env[61868]: _type = "Task" [ 1332.844841] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1332.851810] env[61868]: DEBUG oslo_vmware.api [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316153, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1333.355012] env[61868]: DEBUG oslo_vmware.api [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316153, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.468232} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1333.355012] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] b411db59-3058-4aa8-ac33-d3897c36c123/b411db59-3058-4aa8-ac33-d3897c36c123.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1333.355304] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1333.355354] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5de77858-d1cb-4333-9973-c2ed4fa4dd54 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.361311] env[61868]: DEBUG oslo_vmware.api [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1333.361311] env[61868]: value = "task-1316154" [ 1333.361311] env[61868]: _type = "Task" [ 1333.361311] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1333.368180] env[61868]: DEBUG oslo_vmware.api [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316154, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1333.870632] env[61868]: DEBUG oslo_vmware.api [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316154, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075807} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1333.870894] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1333.871650] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7efd339e-6f97-4623-b48f-19ada3ee6c44 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.892089] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] b411db59-3058-4aa8-ac33-d3897c36c123/b411db59-3058-4aa8-ac33-d3897c36c123.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1333.892269] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-012973f3-5c4c-4728-9164-449a628847a7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.910641] env[61868]: DEBUG oslo_vmware.api [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1333.910641] env[61868]: value = "task-1316155" [ 1333.910641] env[61868]: _type = "Task" [ 1333.910641] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1333.917621] env[61868]: DEBUG oslo_vmware.api [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316155, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1334.420476] env[61868]: DEBUG oslo_vmware.api [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316155, 'name': ReconfigVM_Task, 'duration_secs': 0.277821} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1334.420872] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Reconfigured VM instance instance-0000006f to attach disk [datastore1] b411db59-3058-4aa8-ac33-d3897c36c123/b411db59-3058-4aa8-ac33-d3897c36c123.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1334.421370] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2c5a9bd5-c65b-437c-a54f-25db967b3dcf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.427161] env[61868]: DEBUG oslo_vmware.api [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1334.427161] env[61868]: value = "task-1316156" [ 1334.427161] env[61868]: _type = "Task" [ 1334.427161] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1334.434517] env[61868]: DEBUG oslo_vmware.api [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316156, 'name': Rename_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1334.937704] env[61868]: DEBUG oslo_vmware.api [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316156, 'name': Rename_Task, 'duration_secs': 0.12393} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1334.938108] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1334.938396] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0915149a-9e74-4d3e-bc06-27b31e87aa1e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.944546] env[61868]: DEBUG oslo_vmware.api [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1334.944546] env[61868]: value = "task-1316157" [ 1334.944546] env[61868]: _type = "Task" [ 1334.944546] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1334.952728] env[61868]: DEBUG oslo_vmware.api [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316157, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1335.075176] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1335.075346] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61868) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1335.454796] env[61868]: DEBUG oslo_vmware.api [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316157, 'name': PowerOnVM_Task, 'duration_secs': 0.412182} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1335.455175] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1335.455269] env[61868]: INFO nova.compute.manager [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Took 6.70 seconds to spawn the instance on the hypervisor. [ 1335.455450] env[61868]: DEBUG nova.compute.manager [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1335.456208] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f747104-477a-4aa5-9f3c-0b84b75a9ce1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.971705] env[61868]: INFO nova.compute.manager [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Took 11.85 seconds to build instance. [ 1336.075323] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1336.136576] env[61868]: DEBUG nova.compute.manager [req-05135e59-fedc-4137-a146-718e58107de1 req-f3c79adb-8098-4b21-998b-cbeb34c2587d service nova] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Received event network-changed-22435bb8-30a0-4fbb-b86c-a27082b82e4a {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1336.136796] env[61868]: DEBUG nova.compute.manager [req-05135e59-fedc-4137-a146-718e58107de1 req-f3c79adb-8098-4b21-998b-cbeb34c2587d service nova] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Refreshing instance network info cache due to event network-changed-22435bb8-30a0-4fbb-b86c-a27082b82e4a. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1336.137022] env[61868]: DEBUG oslo_concurrency.lockutils [req-05135e59-fedc-4137-a146-718e58107de1 req-f3c79adb-8098-4b21-998b-cbeb34c2587d service nova] Acquiring lock "refresh_cache-b411db59-3058-4aa8-ac33-d3897c36c123" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1336.137180] env[61868]: DEBUG oslo_concurrency.lockutils [req-05135e59-fedc-4137-a146-718e58107de1 req-f3c79adb-8098-4b21-998b-cbeb34c2587d service nova] Acquired lock "refresh_cache-b411db59-3058-4aa8-ac33-d3897c36c123" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1336.137343] env[61868]: DEBUG nova.network.neutron [req-05135e59-fedc-4137-a146-718e58107de1 req-f3c79adb-8098-4b21-998b-cbeb34c2587d service nova] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Refreshing network info cache for port 22435bb8-30a0-4fbb-b86c-a27082b82e4a {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1336.474426] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ad878a12-4b06-43d3-bb66-bf982ff3eabd tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "b411db59-3058-4aa8-ac33-d3897c36c123" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.361s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1336.855661] env[61868]: DEBUG nova.network.neutron [req-05135e59-fedc-4137-a146-718e58107de1 req-f3c79adb-8098-4b21-998b-cbeb34c2587d service nova] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Updated VIF entry in instance network info cache for port 22435bb8-30a0-4fbb-b86c-a27082b82e4a. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1336.856100] env[61868]: DEBUG nova.network.neutron [req-05135e59-fedc-4137-a146-718e58107de1 req-f3c79adb-8098-4b21-998b-cbeb34c2587d service nova] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Updating instance_info_cache with network_info: [{"id": "22435bb8-30a0-4fbb-b86c-a27082b82e4a", "address": "fa:16:3e:89:f5:a6", "network": {"id": "7e8cbecf-dfdd-43f2-8745-ae1d2d55d309", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-33935587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a7014141b24101ba7920bc83be4a9e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4db2961d-273d-4634-9d06-a94fa9d384fb", "external-id": "nsx-vlan-transportzone-572", "segmentation_id": 572, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22435bb8-30", "ovs_interfaceid": "22435bb8-30a0-4fbb-b86c-a27082b82e4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1337.358501] env[61868]: DEBUG oslo_concurrency.lockutils [req-05135e59-fedc-4137-a146-718e58107de1 req-f3c79adb-8098-4b21-998b-cbeb34c2587d service nova] Releasing lock "refresh_cache-b411db59-3058-4aa8-ac33-d3897c36c123" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1338.075467] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager.update_available_resource {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1338.578818] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1338.579206] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1338.579512] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1338.579759] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61868) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1338.581128] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93610cc2-2b3a-418b-99a7-4d10e8397a93 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.591813] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ea86aa5-0d4c-4d7e-8c4d-e61a54e105e6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.612957] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b70ec2d9-a5c0-4b0a-8d67-8531719d021d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.621580] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e48794e6-09a2-409c-9016-0dcaf1a8e615 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.671775] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180893MB free_disk=134GB free_vcpus=48 pci_devices=None {{(pid=61868) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1338.672169] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1338.672342] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1339.698030] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 59171bf1-f2d4-4ff5-a661-025f5dca7600 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1339.698030] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 32abbff4-5d9f-4c4c-8270-e3e5d383825e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1339.698374] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 77e19bcd-5963-4c48-bcf2-da5c92b142ff actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1339.698374] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance b411db59-3058-4aa8-ac33-d3897c36c123 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1339.698533] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=61868) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1339.698586] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1344MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=61868) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1339.753693] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9813d28-37c7-49bd-9c6c-5a2525546f93 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.760936] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad39105c-6b13-4b11-8f2c-77d7b6606991 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.788934] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e43dac6f-d241-4a82-bdbd-907aea8859d1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.795976] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d62b045-38e0-4279-ade5-0fe920cc865e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.808351] env[61868]: DEBUG nova.compute.provider_tree [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1340.311850] env[61868]: DEBUG nova.scheduler.client.report [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1340.817709] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61868) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1340.818109] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.146s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1341.818634] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1341.818930] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1341.819074] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Starting heal instance info cache {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1342.853200] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "refresh_cache-32abbff4-5d9f-4c4c-8270-e3e5d383825e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1342.853416] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquired lock "refresh_cache-32abbff4-5d9f-4c4c-8270-e3e5d383825e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1342.853501] env[61868]: DEBUG nova.network.neutron [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Forcefully refreshing network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1344.038693] env[61868]: DEBUG nova.network.neutron [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Updating instance_info_cache with network_info: [{"id": "6840da6b-2963-47e2-a5da-cf0e9d1d0fb8", "address": "fa:16:3e:c4:fa:ea", "network": {"id": "16b3638c-1129-49b8-9f22-636ff7db7386", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1544179589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bdab2203834c4ee5bab11aca70e48cc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d298db54-f13d-4bf6-b6c2-755074b3047f", "external-id": "nsx-vlan-transportzone-631", "segmentation_id": 631, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6840da6b-29", "ovs_interfaceid": "6840da6b-2963-47e2-a5da-cf0e9d1d0fb8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1344.541505] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Releasing lock "refresh_cache-32abbff4-5d9f-4c4c-8270-e3e5d383825e" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1344.541737] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Updated the network info_cache for instance {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1344.541935] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1344.542118] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1344.542272] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1344.542415] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1348.793771] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1354.652054] env[61868]: DEBUG oslo_concurrency.lockutils [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "77e19bcd-5963-4c48-bcf2-da5c92b142ff" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1354.652054] env[61868]: DEBUG oslo_concurrency.lockutils [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "77e19bcd-5963-4c48-bcf2-da5c92b142ff" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1354.652054] env[61868]: DEBUG oslo_concurrency.lockutils [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "77e19bcd-5963-4c48-bcf2-da5c92b142ff-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1354.652054] env[61868]: DEBUG oslo_concurrency.lockutils [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "77e19bcd-5963-4c48-bcf2-da5c92b142ff-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1354.652054] env[61868]: DEBUG oslo_concurrency.lockutils [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "77e19bcd-5963-4c48-bcf2-da5c92b142ff-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1354.655356] env[61868]: INFO nova.compute.manager [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Terminating instance [ 1354.657329] env[61868]: DEBUG nova.compute.manager [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1354.657546] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1354.657784] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-28dc9fa6-c528-49b1-a70f-a47bc9cc67ac {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.665153] env[61868]: DEBUG oslo_vmware.api [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1354.665153] env[61868]: value = "task-1316158" [ 1354.665153] env[61868]: _type = "Task" [ 1354.665153] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1354.672678] env[61868]: DEBUG oslo_vmware.api [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316158, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1355.175633] env[61868]: DEBUG oslo_vmware.api [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316158, 'name': PowerOffVM_Task, 'duration_secs': 0.156558} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1355.175878] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1355.176094] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Volume detach. Driver type: vmdk {{(pid=61868) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1355.176297] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281647', 'volume_id': 'ec4ed682-190b-4dbe-b2f8-1b4e3e82fb91', 'name': 'volume-ec4ed682-190b-4dbe-b2f8-1b4e3e82fb91', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '77e19bcd-5963-4c48-bcf2-da5c92b142ff', 'attached_at': '2024-10-18T17:32:04.000000', 'detached_at': '', 'volume_id': 'ec4ed682-190b-4dbe-b2f8-1b4e3e82fb91', 'serial': 'ec4ed682-190b-4dbe-b2f8-1b4e3e82fb91'} {{(pid=61868) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1355.177024] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52f8993f-5e92-449b-af66-4421d63e26d4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1355.194307] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f86cdf5-d575-444d-b857-f649fa614f97 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1355.200409] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84828ee2-a7a1-42a6-98c9-2dc655404352 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1355.216996] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e8b1643-98f2-4b7b-9a99-cacfb0d06727 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1355.230369] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] The volume has not been displaced from its original location: [datastore2] volume-ec4ed682-190b-4dbe-b2f8-1b4e3e82fb91/volume-ec4ed682-190b-4dbe-b2f8-1b4e3e82fb91.vmdk. No consolidation needed. {{(pid=61868) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1355.235500] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Reconfiguring VM instance instance-0000006e to detach disk 2000 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1355.235737] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8553099b-be06-4be9-ae42-cabecf1bfe78 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1355.252025] env[61868]: DEBUG oslo_vmware.api [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1355.252025] env[61868]: value = "task-1316159" [ 1355.252025] env[61868]: _type = "Task" [ 1355.252025] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1355.258697] env[61868]: DEBUG oslo_vmware.api [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316159, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1355.761458] env[61868]: DEBUG oslo_vmware.api [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316159, 'name': ReconfigVM_Task, 'duration_secs': 0.150055} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1355.761842] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Reconfigured VM instance instance-0000006e to detach disk 2000 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1355.766285] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-08c005c2-0fc5-440a-9573-bea8268ce1c1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1355.780856] env[61868]: DEBUG oslo_vmware.api [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1355.780856] env[61868]: value = "task-1316160" [ 1355.780856] env[61868]: _type = "Task" [ 1355.780856] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1355.788148] env[61868]: DEBUG oslo_vmware.api [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316160, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1356.290279] env[61868]: DEBUG oslo_vmware.api [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316160, 'name': ReconfigVM_Task, 'duration_secs': 0.147573} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1356.290584] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281647', 'volume_id': 'ec4ed682-190b-4dbe-b2f8-1b4e3e82fb91', 'name': 'volume-ec4ed682-190b-4dbe-b2f8-1b4e3e82fb91', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '77e19bcd-5963-4c48-bcf2-da5c92b142ff', 'attached_at': '2024-10-18T17:32:04.000000', 'detached_at': '', 'volume_id': 'ec4ed682-190b-4dbe-b2f8-1b4e3e82fb91', 'serial': 'ec4ed682-190b-4dbe-b2f8-1b4e3e82fb91'} {{(pid=61868) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1356.290845] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1356.291888] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bdb4ce5-38a4-4433-bf33-2c29dae74635 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.297840] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1356.298075] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2cf4d210-20d1-409e-a304-22569d071fbb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.357962] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1356.358233] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Deleting contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1356.358426] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Deleting the datastore file [datastore2] 77e19bcd-5963-4c48-bcf2-da5c92b142ff {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1356.358685] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cc7473b7-4f61-4d2a-967b-899e150e2bfb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.365632] env[61868]: DEBUG oslo_vmware.api [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1356.365632] env[61868]: value = "task-1316162" [ 1356.365632] env[61868]: _type = "Task" [ 1356.365632] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1356.372839] env[61868]: DEBUG oslo_vmware.api [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316162, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1356.875911] env[61868]: DEBUG oslo_vmware.api [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316162, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.080498} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1356.876341] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1356.876341] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Deleted contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1356.876525] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1356.876696] env[61868]: INFO nova.compute.manager [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Took 2.22 seconds to destroy the instance on the hypervisor. [ 1356.876931] env[61868]: DEBUG oslo.service.loopingcall [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1356.877161] env[61868]: DEBUG nova.compute.manager [-] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1356.877255] env[61868]: DEBUG nova.network.neutron [-] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1357.293017] env[61868]: DEBUG nova.compute.manager [req-8ffd4c8a-a4d0-46a0-8f5e-320669d3cd64 req-b7523033-5a45-40df-b088-737215902d58 service nova] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Received event network-vif-deleted-eab3dce4-b36e-4ab7-bc96-27545c10e311 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1357.293239] env[61868]: INFO nova.compute.manager [req-8ffd4c8a-a4d0-46a0-8f5e-320669d3cd64 req-b7523033-5a45-40df-b088-737215902d58 service nova] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Neutron deleted interface eab3dce4-b36e-4ab7-bc96-27545c10e311; detaching it from the instance and deleting it from the info cache [ 1357.293483] env[61868]: DEBUG nova.network.neutron [req-8ffd4c8a-a4d0-46a0-8f5e-320669d3cd64 req-b7523033-5a45-40df-b088-737215902d58 service nova] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1357.768501] env[61868]: DEBUG nova.network.neutron [-] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1357.796446] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ab4e3529-e0a9-48a1-bcd9-9516f782159e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1357.805683] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4226332f-ef74-4cb7-b02b-b122fad98dea {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1357.831594] env[61868]: DEBUG nova.compute.manager [req-8ffd4c8a-a4d0-46a0-8f5e-320669d3cd64 req-b7523033-5a45-40df-b088-737215902d58 service nova] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Detach interface failed, port_id=eab3dce4-b36e-4ab7-bc96-27545c10e311, reason: Instance 77e19bcd-5963-4c48-bcf2-da5c92b142ff could not be found. {{(pid=61868) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1358.270869] env[61868]: INFO nova.compute.manager [-] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Took 1.39 seconds to deallocate network for instance. [ 1358.816972] env[61868]: INFO nova.compute.manager [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Took 0.55 seconds to detach 1 volumes for instance. [ 1358.819168] env[61868]: DEBUG nova.compute.manager [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Deleting volume: ec4ed682-190b-4dbe-b2f8-1b4e3e82fb91 {{(pid=61868) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1359.356731] env[61868]: DEBUG oslo_concurrency.lockutils [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1359.357038] env[61868]: DEBUG oslo_concurrency.lockutils [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1359.357282] env[61868]: DEBUG nova.objects.instance [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lazy-loading 'resources' on Instance uuid 77e19bcd-5963-4c48-bcf2-da5c92b142ff {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1359.919943] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1079f2e6-b3a8-41ab-94ac-b8babc2427d0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.927770] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-286a5605-51c3-4f04-9613-1f4f1a66f08a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.957745] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4810df29-8b32-42f9-b6fb-e928fbb893e5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.965198] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45506b36-31d9-4296-ad6f-fc4172ff1a78 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.978118] env[61868]: DEBUG nova.compute.provider_tree [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1360.481517] env[61868]: DEBUG nova.scheduler.client.report [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1360.986668] env[61868]: DEBUG oslo_concurrency.lockutils [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.629s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1361.004034] env[61868]: INFO nova.scheduler.client.report [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Deleted allocations for instance 77e19bcd-5963-4c48-bcf2-da5c92b142ff [ 1361.511886] env[61868]: DEBUG oslo_concurrency.lockutils [None req-73792791-f736-46e8-a79f-5f732c1a6cf2 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "77e19bcd-5963-4c48-bcf2-da5c92b142ff" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.860s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1362.308856] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5f8d4493-5f96-4575-93a9-66533017c7c7 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "32abbff4-5d9f-4c4c-8270-e3e5d383825e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1362.309190] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5f8d4493-5f96-4575-93a9-66533017c7c7 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "32abbff4-5d9f-4c4c-8270-e3e5d383825e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1362.309423] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5f8d4493-5f96-4575-93a9-66533017c7c7 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "32abbff4-5d9f-4c4c-8270-e3e5d383825e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1362.309617] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5f8d4493-5f96-4575-93a9-66533017c7c7 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "32abbff4-5d9f-4c4c-8270-e3e5d383825e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1362.309794] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5f8d4493-5f96-4575-93a9-66533017c7c7 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "32abbff4-5d9f-4c4c-8270-e3e5d383825e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1362.311966] env[61868]: INFO nova.compute.manager [None req-5f8d4493-5f96-4575-93a9-66533017c7c7 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Terminating instance [ 1362.313666] env[61868]: DEBUG nova.compute.manager [None req-5f8d4493-5f96-4575-93a9-66533017c7c7 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1362.313862] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-5f8d4493-5f96-4575-93a9-66533017c7c7 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1362.314682] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af016331-5246-4367-8eaa-a182a69fdd8f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.322031] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f8d4493-5f96-4575-93a9-66533017c7c7 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1362.322257] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3bd54649-5a30-4671-bbd8-96d5531da1a3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.328569] env[61868]: DEBUG oslo_vmware.api [None req-5f8d4493-5f96-4575-93a9-66533017c7c7 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1362.328569] env[61868]: value = "task-1316164" [ 1362.328569] env[61868]: _type = "Task" [ 1362.328569] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1362.335726] env[61868]: DEBUG oslo_vmware.api [None req-5f8d4493-5f96-4575-93a9-66533017c7c7 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316164, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1362.838589] env[61868]: DEBUG oslo_vmware.api [None req-5f8d4493-5f96-4575-93a9-66533017c7c7 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316164, 'name': PowerOffVM_Task, 'duration_secs': 0.174243} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1362.838965] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f8d4493-5f96-4575-93a9-66533017c7c7 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1362.838965] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-5f8d4493-5f96-4575-93a9-66533017c7c7 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1362.839209] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-222a61fc-daa2-4282-921e-8ec04a6852ac {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1363.563722] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-5f8d4493-5f96-4575-93a9-66533017c7c7 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1363.563985] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-5f8d4493-5f96-4575-93a9-66533017c7c7 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1363.564150] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f8d4493-5f96-4575-93a9-66533017c7c7 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Deleting the datastore file [datastore1] 32abbff4-5d9f-4c4c-8270-e3e5d383825e {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1363.564414] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5a3ae998-7d77-41a2-8f48-14d8015bbb44 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1363.572332] env[61868]: DEBUG oslo_vmware.api [None req-5f8d4493-5f96-4575-93a9-66533017c7c7 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1363.572332] env[61868]: value = "task-1316166" [ 1363.572332] env[61868]: _type = "Task" [ 1363.572332] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1363.579468] env[61868]: DEBUG oslo_vmware.api [None req-5f8d4493-5f96-4575-93a9-66533017c7c7 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316166, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1364.081728] env[61868]: DEBUG oslo_vmware.api [None req-5f8d4493-5f96-4575-93a9-66533017c7c7 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316166, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147476} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1364.082065] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f8d4493-5f96-4575-93a9-66533017c7c7 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1364.082168] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-5f8d4493-5f96-4575-93a9-66533017c7c7 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1364.082352] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-5f8d4493-5f96-4575-93a9-66533017c7c7 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1364.082529] env[61868]: INFO nova.compute.manager [None req-5f8d4493-5f96-4575-93a9-66533017c7c7 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Took 1.77 seconds to destroy the instance on the hypervisor. [ 1364.082771] env[61868]: DEBUG oslo.service.loopingcall [None req-5f8d4493-5f96-4575-93a9-66533017c7c7 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1364.082962] env[61868]: DEBUG nova.compute.manager [-] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1364.083068] env[61868]: DEBUG nova.network.neutron [-] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1364.318979] env[61868]: DEBUG nova.compute.manager [req-bce93df9-35e9-428f-a5c5-3802bd5b45dc req-f3bf12f9-590d-428e-8104-5b955227658d service nova] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Received event network-vif-deleted-6840da6b-2963-47e2-a5da-cf0e9d1d0fb8 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1364.318979] env[61868]: INFO nova.compute.manager [req-bce93df9-35e9-428f-a5c5-3802bd5b45dc req-f3bf12f9-590d-428e-8104-5b955227658d service nova] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Neutron deleted interface 6840da6b-2963-47e2-a5da-cf0e9d1d0fb8; detaching it from the instance and deleting it from the info cache [ 1364.318979] env[61868]: DEBUG nova.network.neutron [req-bce93df9-35e9-428f-a5c5-3802bd5b45dc req-f3bf12f9-590d-428e-8104-5b955227658d service nova] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1364.798534] env[61868]: DEBUG nova.network.neutron [-] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1364.821359] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dbfed673-f24c-49cb-9edc-53dbbfe691fb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.830429] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ec70fd9-bf3e-42d1-a618-dcd4acbe53bb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.853276] env[61868]: DEBUG nova.compute.manager [req-bce93df9-35e9-428f-a5c5-3802bd5b45dc req-f3bf12f9-590d-428e-8104-5b955227658d service nova] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Detach interface failed, port_id=6840da6b-2963-47e2-a5da-cf0e9d1d0fb8, reason: Instance 32abbff4-5d9f-4c4c-8270-e3e5d383825e could not be found. {{(pid=61868) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1365.301740] env[61868]: INFO nova.compute.manager [-] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Took 1.22 seconds to deallocate network for instance. [ 1365.809027] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5f8d4493-5f96-4575-93a9-66533017c7c7 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1365.809261] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5f8d4493-5f96-4575-93a9-66533017c7c7 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1365.809501] env[61868]: DEBUG nova.objects.instance [None req-5f8d4493-5f96-4575-93a9-66533017c7c7 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lazy-loading 'resources' on Instance uuid 32abbff4-5d9f-4c4c-8270-e3e5d383825e {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1366.363370] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3e9c032-5634-4039-89a3-540b4344f1aa {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1366.370681] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-967ec44c-0c0b-4ebc-bd8a-97a84d145153 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1366.400169] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0123177a-df0a-40c9-afbd-79698e51bbc3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1366.406735] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53868779-bb22-4416-9c5b-d465d8fc93ba {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1366.419261] env[61868]: DEBUG nova.compute.provider_tree [None req-5f8d4493-5f96-4575-93a9-66533017c7c7 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1366.921964] env[61868]: DEBUG nova.scheduler.client.report [None req-5f8d4493-5f96-4575-93a9-66533017c7c7 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1367.427877] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5f8d4493-5f96-4575-93a9-66533017c7c7 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.618s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1367.450270] env[61868]: INFO nova.scheduler.client.report [None req-5f8d4493-5f96-4575-93a9-66533017c7c7 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Deleted allocations for instance 32abbff4-5d9f-4c4c-8270-e3e5d383825e [ 1367.958267] env[61868]: DEBUG oslo_concurrency.lockutils [None req-5f8d4493-5f96-4575-93a9-66533017c7c7 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "32abbff4-5d9f-4c4c-8270-e3e5d383825e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.648s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1370.971058] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d16cdf31-7501-4f07-b7c6-9fbe8e61cf55 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "59171bf1-f2d4-4ff5-a661-025f5dca7600" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1370.971450] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d16cdf31-7501-4f07-b7c6-9fbe8e61cf55 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "59171bf1-f2d4-4ff5-a661-025f5dca7600" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1370.971450] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d16cdf31-7501-4f07-b7c6-9fbe8e61cf55 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "59171bf1-f2d4-4ff5-a661-025f5dca7600-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1370.971527] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d16cdf31-7501-4f07-b7c6-9fbe8e61cf55 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "59171bf1-f2d4-4ff5-a661-025f5dca7600-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1370.971676] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d16cdf31-7501-4f07-b7c6-9fbe8e61cf55 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "59171bf1-f2d4-4ff5-a661-025f5dca7600-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1370.973749] env[61868]: INFO nova.compute.manager [None req-d16cdf31-7501-4f07-b7c6-9fbe8e61cf55 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Terminating instance [ 1370.975467] env[61868]: DEBUG nova.compute.manager [None req-d16cdf31-7501-4f07-b7c6-9fbe8e61cf55 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1370.975661] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-d16cdf31-7501-4f07-b7c6-9fbe8e61cf55 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1370.976489] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0fee5e6-783d-40ae-b09a-9b4f288cfdd3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1370.984071] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-d16cdf31-7501-4f07-b7c6-9fbe8e61cf55 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1370.984293] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e7e07e9b-f023-4b99-b4f2-5dd063af7928 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1370.991010] env[61868]: DEBUG oslo_vmware.api [None req-d16cdf31-7501-4f07-b7c6-9fbe8e61cf55 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1370.991010] env[61868]: value = "task-1316168" [ 1370.991010] env[61868]: _type = "Task" [ 1370.991010] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1370.998152] env[61868]: DEBUG oslo_vmware.api [None req-d16cdf31-7501-4f07-b7c6-9fbe8e61cf55 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316168, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1371.501364] env[61868]: DEBUG oslo_vmware.api [None req-d16cdf31-7501-4f07-b7c6-9fbe8e61cf55 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316168, 'name': PowerOffVM_Task, 'duration_secs': 0.179848} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1371.501643] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-d16cdf31-7501-4f07-b7c6-9fbe8e61cf55 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1371.501793] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-d16cdf31-7501-4f07-b7c6-9fbe8e61cf55 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1371.502039] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2f66797e-2e03-43f3-9546-5ff94684ef3e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1371.563191] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-d16cdf31-7501-4f07-b7c6-9fbe8e61cf55 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1371.563423] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-d16cdf31-7501-4f07-b7c6-9fbe8e61cf55 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Deleting contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1371.563608] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-d16cdf31-7501-4f07-b7c6-9fbe8e61cf55 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Deleting the datastore file [datastore2] 59171bf1-f2d4-4ff5-a661-025f5dca7600 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1371.563863] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-34cbe9c4-20b7-4479-bea4-a6bf6529342d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1371.570239] env[61868]: DEBUG oslo_vmware.api [None req-d16cdf31-7501-4f07-b7c6-9fbe8e61cf55 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for the task: (returnval){ [ 1371.570239] env[61868]: value = "task-1316170" [ 1371.570239] env[61868]: _type = "Task" [ 1371.570239] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1371.577722] env[61868]: DEBUG oslo_vmware.api [None req-d16cdf31-7501-4f07-b7c6-9fbe8e61cf55 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316170, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1372.080249] env[61868]: DEBUG oslo_vmware.api [None req-d16cdf31-7501-4f07-b7c6-9fbe8e61cf55 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Task: {'id': task-1316170, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.130895} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1372.080628] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-d16cdf31-7501-4f07-b7c6-9fbe8e61cf55 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1372.080676] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-d16cdf31-7501-4f07-b7c6-9fbe8e61cf55 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Deleted contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1372.080854] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-d16cdf31-7501-4f07-b7c6-9fbe8e61cf55 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1372.081041] env[61868]: INFO nova.compute.manager [None req-d16cdf31-7501-4f07-b7c6-9fbe8e61cf55 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1372.081302] env[61868]: DEBUG oslo.service.loopingcall [None req-d16cdf31-7501-4f07-b7c6-9fbe8e61cf55 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1372.081502] env[61868]: DEBUG nova.compute.manager [-] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1372.081596] env[61868]: DEBUG nova.network.neutron [-] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1372.316142] env[61868]: DEBUG nova.compute.manager [req-bb84933c-bd0b-4ef9-88ad-c1378e3a684d req-061ecca6-fa99-41ad-b68e-497eb49a1fb1 service nova] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Received event network-vif-deleted-8f9ad20f-f367-4826-8aae-00a1192d0bc8 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1372.316268] env[61868]: INFO nova.compute.manager [req-bb84933c-bd0b-4ef9-88ad-c1378e3a684d req-061ecca6-fa99-41ad-b68e-497eb49a1fb1 service nova] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Neutron deleted interface 8f9ad20f-f367-4826-8aae-00a1192d0bc8; detaching it from the instance and deleting it from the info cache [ 1372.316412] env[61868]: DEBUG nova.network.neutron [req-bb84933c-bd0b-4ef9-88ad-c1378e3a684d req-061ecca6-fa99-41ad-b68e-497eb49a1fb1 service nova] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1372.796772] env[61868]: DEBUG nova.network.neutron [-] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1372.819125] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e02f229e-1e43-4849-a984-0aa7d1f3214a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1372.828995] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-563efb3b-e290-4ee3-9963-407e8d2d15ec {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1372.850706] env[61868]: DEBUG nova.compute.manager [req-bb84933c-bd0b-4ef9-88ad-c1378e3a684d req-061ecca6-fa99-41ad-b68e-497eb49a1fb1 service nova] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Detach interface failed, port_id=8f9ad20f-f367-4826-8aae-00a1192d0bc8, reason: Instance 59171bf1-f2d4-4ff5-a661-025f5dca7600 could not be found. {{(pid=61868) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1373.299530] env[61868]: INFO nova.compute.manager [-] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Took 1.22 seconds to deallocate network for instance. [ 1373.806615] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d16cdf31-7501-4f07-b7c6-9fbe8e61cf55 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1373.806615] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d16cdf31-7501-4f07-b7c6-9fbe8e61cf55 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1373.806856] env[61868]: DEBUG nova.objects.instance [None req-d16cdf31-7501-4f07-b7c6-9fbe8e61cf55 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lazy-loading 'resources' on Instance uuid 59171bf1-f2d4-4ff5-a661-025f5dca7600 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1373.886284] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "b411db59-3058-4aa8-ac33-d3897c36c123" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1373.886534] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "b411db59-3058-4aa8-ac33-d3897c36c123" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1373.886719] env[61868]: INFO nova.compute.manager [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Shelving [ 1374.353728] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d20eb8af-48ea-40d9-9d08-7b59126b1f6b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1374.360976] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e8da3e8-38f3-402b-992b-9421baf06cd2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1374.391652] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b06ed5b-d553-41b8-a0e1-945d3c74a761 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1374.396274] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1374.396531] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f95a0f3a-366b-4663-ac86-d59afd8cd499 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1374.401074] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b8c9269-1b49-4195-81e3-746b658ee095 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1374.405851] env[61868]: DEBUG oslo_vmware.api [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1374.405851] env[61868]: value = "task-1316171" [ 1374.405851] env[61868]: _type = "Task" [ 1374.405851] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1374.416947] env[61868]: DEBUG nova.compute.provider_tree [None req-d16cdf31-7501-4f07-b7c6-9fbe8e61cf55 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1374.422488] env[61868]: DEBUG oslo_vmware.api [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316171, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1374.916281] env[61868]: DEBUG oslo_vmware.api [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316171, 'name': PowerOffVM_Task, 'duration_secs': 0.190913} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1374.916569] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1374.917456] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f240e751-4079-4428-9132-bf984c244d13 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1374.920506] env[61868]: DEBUG nova.scheduler.client.report [None req-d16cdf31-7501-4f07-b7c6-9fbe8e61cf55 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1374.940902] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-519000a6-6b22-4f86-80c8-050aa216c110 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1375.783941] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d16cdf31-7501-4f07-b7c6-9fbe8e61cf55 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.977s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1375.786850] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Creating Snapshot of the VM instance {{(pid=61868) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1375.787353] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-fd4e9e91-516e-4a0e-89af-33073782a989 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1375.794533] env[61868]: DEBUG oslo_vmware.api [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1375.794533] env[61868]: value = "task-1316172" [ 1375.794533] env[61868]: _type = "Task" [ 1375.794533] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1375.801950] env[61868]: DEBUG oslo_vmware.api [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316172, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1375.804845] env[61868]: INFO nova.scheduler.client.report [None req-d16cdf31-7501-4f07-b7c6-9fbe8e61cf55 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Deleted allocations for instance 59171bf1-f2d4-4ff5-a661-025f5dca7600 [ 1376.304016] env[61868]: DEBUG oslo_vmware.api [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316172, 'name': CreateSnapshot_Task, 'duration_secs': 0.431878} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1376.304297] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Created Snapshot of the VM instance {{(pid=61868) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1376.305065] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b11c873-639b-4c8c-ac0c-fad1676509bb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1376.316546] env[61868]: DEBUG oslo_concurrency.lockutils [None req-d16cdf31-7501-4f07-b7c6-9fbe8e61cf55 tempest-ServerActionsTestOtherA-1056544490 tempest-ServerActionsTestOtherA-1056544490-project-member] Lock "59171bf1-f2d4-4ff5-a661-025f5dca7600" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.345s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1376.824685] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Creating linked-clone VM from snapshot {{(pid=61868) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1376.825389] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-0ffcfe68-f1d3-4313-8e06-9db96dc0a917 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1376.834237] env[61868]: DEBUG oslo_vmware.api [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1376.834237] env[61868]: value = "task-1316173" [ 1376.834237] env[61868]: _type = "Task" [ 1376.834237] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1376.842107] env[61868]: DEBUG oslo_vmware.api [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316173, 'name': CloneVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1377.344102] env[61868]: DEBUG oslo_vmware.api [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316173, 'name': CloneVM_Task} progress is 94%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1377.844648] env[61868]: DEBUG oslo_vmware.api [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316173, 'name': CloneVM_Task} progress is 95%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1378.345141] env[61868]: DEBUG oslo_vmware.api [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316173, 'name': CloneVM_Task, 'duration_secs': 1.100265} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1378.345428] env[61868]: INFO nova.virt.vmwareapi.vmops [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Created linked-clone VM from snapshot [ 1378.346140] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adb0f066-49fc-4750-b8bb-a6ffdd3bc558 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1378.352735] env[61868]: DEBUG nova.virt.vmwareapi.images [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Uploading image 6308358b-8a38-4242-ad10-cd94d2d8a7fb {{(pid=61868) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1378.377415] env[61868]: DEBUG oslo_vmware.rw_handles [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1378.377415] env[61868]: value = "vm-281651" [ 1378.377415] env[61868]: _type = "VirtualMachine" [ 1378.377415] env[61868]: }. {{(pid=61868) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1378.377672] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-9e3d9644-6270-4c3f-b6c7-62668a596364 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1378.384942] env[61868]: DEBUG oslo_vmware.rw_handles [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lease: (returnval){ [ 1378.384942] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e28286-c906-c7d2-2ac6-541f741592f3" [ 1378.384942] env[61868]: _type = "HttpNfcLease" [ 1378.384942] env[61868]: } obtained for exporting VM: (result){ [ 1378.384942] env[61868]: value = "vm-281651" [ 1378.384942] env[61868]: _type = "VirtualMachine" [ 1378.384942] env[61868]: }. {{(pid=61868) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1378.385314] env[61868]: DEBUG oslo_vmware.api [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the lease: (returnval){ [ 1378.385314] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e28286-c906-c7d2-2ac6-541f741592f3" [ 1378.385314] env[61868]: _type = "HttpNfcLease" [ 1378.385314] env[61868]: } to be ready. {{(pid=61868) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1378.391314] env[61868]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1378.391314] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e28286-c906-c7d2-2ac6-541f741592f3" [ 1378.391314] env[61868]: _type = "HttpNfcLease" [ 1378.391314] env[61868]: } is initializing. {{(pid=61868) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1378.892677] env[61868]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1378.892677] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e28286-c906-c7d2-2ac6-541f741592f3" [ 1378.892677] env[61868]: _type = "HttpNfcLease" [ 1378.892677] env[61868]: } is ready. {{(pid=61868) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1378.893021] env[61868]: DEBUG oslo_vmware.rw_handles [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1378.893021] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e28286-c906-c7d2-2ac6-541f741592f3" [ 1378.893021] env[61868]: _type = "HttpNfcLease" [ 1378.893021] env[61868]: }. {{(pid=61868) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1378.894025] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-745fda86-aea1-4b4c-99be-a93ec6126bde {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1378.900498] env[61868]: DEBUG oslo_vmware.rw_handles [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f49862-d63e-eb59-7f21-288e22b279ce/disk-0.vmdk from lease info. {{(pid=61868) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1378.900672] env[61868]: DEBUG oslo_vmware.rw_handles [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f49862-d63e-eb59-7f21-288e22b279ce/disk-0.vmdk for reading. {{(pid=61868) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1378.990080] env[61868]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-0b45aec1-72ef-4f0e-a9fb-38319b95efcb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1381.178483] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquiring lock "28a14ebd-6003-4e20-b556-2a3b939f15c8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1381.179352] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lock "28a14ebd-6003-4e20-b556-2a3b939f15c8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1381.506948] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquiring lock "6ce00281-a201-445a-bb75-aac07fd481aa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1381.507319] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lock "6ce00281-a201-445a-bb75-aac07fd481aa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1381.681884] env[61868]: DEBUG nova.compute.manager [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1382.013027] env[61868]: DEBUG nova.compute.manager [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1382.206592] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1382.207173] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1382.208683] env[61868]: INFO nova.compute.claims [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1382.530157] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1383.269433] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd8148b2-3432-4597-8164-325d3a44f8bf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1383.277282] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4531ece0-c71d-4f54-a1ee-2c59d10e942b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1383.306930] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63fc5386-5eb2-496c-b54c-1ec3d6b29459 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1383.314056] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86bb8684-646e-40f8-8b9e-ff50b2c98e06 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1383.326813] env[61868]: DEBUG nova.compute.provider_tree [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1383.829920] env[61868]: DEBUG nova.scheduler.client.report [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1384.335494] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.128s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1384.336058] env[61868]: DEBUG nova.compute.manager [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1384.339419] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.809s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1384.340794] env[61868]: INFO nova.compute.claims [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1384.844586] env[61868]: DEBUG nova.compute.utils [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1384.848854] env[61868]: DEBUG nova.compute.manager [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1384.849142] env[61868]: DEBUG nova.network.neutron [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1384.890080] env[61868]: DEBUG nova.policy [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4dbe667c15f14d13b6ef69bfe770ac70', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b3904d578a0e47d58eae75b1d4fbf41c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 1385.167164] env[61868]: DEBUG nova.network.neutron [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Successfully created port: b1bf2dec-5f14-461f-aa54-f6d67bc0dcb5 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1385.353231] env[61868]: DEBUG nova.compute.manager [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1385.405369] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f93f53ca-36ee-45e1-96c8-ba536cc721c7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1385.413365] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5470cb00-67be-4fb5-909d-293f67feb1f6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1385.445819] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7c2401e-dbef-4e7d-9f85-dde33e3d5beb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1385.453969] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ddf502f-c9ac-4268-95a2-044ebf0e7aec {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1385.468495] env[61868]: DEBUG nova.compute.provider_tree [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1385.742746] env[61868]: DEBUG oslo_vmware.rw_handles [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f49862-d63e-eb59-7f21-288e22b279ce/disk-0.vmdk. {{(pid=61868) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1385.743197] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91d6c732-5af8-4426-b6fd-e20a9b3c83db {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1385.749547] env[61868]: DEBUG oslo_vmware.rw_handles [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f49862-d63e-eb59-7f21-288e22b279ce/disk-0.vmdk is in state: ready. {{(pid=61868) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1385.749716] env[61868]: ERROR oslo_vmware.rw_handles [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f49862-d63e-eb59-7f21-288e22b279ce/disk-0.vmdk due to incomplete transfer. [ 1385.749925] env[61868]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-e6ff55f4-fef8-45aa-a047-a65928c34a2b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1385.757377] env[61868]: DEBUG oslo_vmware.rw_handles [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f49862-d63e-eb59-7f21-288e22b279ce/disk-0.vmdk. {{(pid=61868) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1385.757611] env[61868]: DEBUG nova.virt.vmwareapi.images [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Uploaded image 6308358b-8a38-4242-ad10-cd94d2d8a7fb to the Glance image server {{(pid=61868) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1385.759786] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Destroying the VM {{(pid=61868) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1385.760029] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-32cb1cce-1c84-4afa-9136-1fe695acc4e8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1385.765953] env[61868]: DEBUG oslo_vmware.api [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1385.765953] env[61868]: value = "task-1316175" [ 1385.765953] env[61868]: _type = "Task" [ 1385.765953] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1385.773830] env[61868]: DEBUG oslo_vmware.api [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316175, 'name': Destroy_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1385.971618] env[61868]: DEBUG nova.scheduler.client.report [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1386.275508] env[61868]: DEBUG oslo_vmware.api [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316175, 'name': Destroy_Task} progress is 100%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1386.366023] env[61868]: DEBUG nova.compute.manager [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1386.388095] env[61868]: DEBUG nova.virt.hardware [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1386.388345] env[61868]: DEBUG nova.virt.hardware [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1386.388529] env[61868]: DEBUG nova.virt.hardware [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1386.388716] env[61868]: DEBUG nova.virt.hardware [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1386.388907] env[61868]: DEBUG nova.virt.hardware [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1386.389086] env[61868]: DEBUG nova.virt.hardware [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1386.389300] env[61868]: DEBUG nova.virt.hardware [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1386.389463] env[61868]: DEBUG nova.virt.hardware [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1386.389635] env[61868]: DEBUG nova.virt.hardware [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1386.389797] env[61868]: DEBUG nova.virt.hardware [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1386.389972] env[61868]: DEBUG nova.virt.hardware [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1386.390823] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a380393-ec8c-44b0-87f3-b9636be4a425 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1386.398296] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69a3e0e9-3650-4b06-8e49-31c24305a5e6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1386.476828] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.137s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1386.477449] env[61868]: DEBUG nova.compute.manager [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1386.560630] env[61868]: DEBUG nova.compute.manager [req-15b00dd4-6bdf-450b-9cfe-e3cb1def5ab0 req-46e33523-0bee-456c-90d5-2445785d70d6 service nova] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Received event network-vif-plugged-b1bf2dec-5f14-461f-aa54-f6d67bc0dcb5 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1386.560911] env[61868]: DEBUG oslo_concurrency.lockutils [req-15b00dd4-6bdf-450b-9cfe-e3cb1def5ab0 req-46e33523-0bee-456c-90d5-2445785d70d6 service nova] Acquiring lock "28a14ebd-6003-4e20-b556-2a3b939f15c8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1386.561185] env[61868]: DEBUG oslo_concurrency.lockutils [req-15b00dd4-6bdf-450b-9cfe-e3cb1def5ab0 req-46e33523-0bee-456c-90d5-2445785d70d6 service nova] Lock "28a14ebd-6003-4e20-b556-2a3b939f15c8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1386.561365] env[61868]: DEBUG oslo_concurrency.lockutils [req-15b00dd4-6bdf-450b-9cfe-e3cb1def5ab0 req-46e33523-0bee-456c-90d5-2445785d70d6 service nova] Lock "28a14ebd-6003-4e20-b556-2a3b939f15c8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1386.561585] env[61868]: DEBUG nova.compute.manager [req-15b00dd4-6bdf-450b-9cfe-e3cb1def5ab0 req-46e33523-0bee-456c-90d5-2445785d70d6 service nova] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] No waiting events found dispatching network-vif-plugged-b1bf2dec-5f14-461f-aa54-f6d67bc0dcb5 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1386.561716] env[61868]: WARNING nova.compute.manager [req-15b00dd4-6bdf-450b-9cfe-e3cb1def5ab0 req-46e33523-0bee-456c-90d5-2445785d70d6 service nova] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Received unexpected event network-vif-plugged-b1bf2dec-5f14-461f-aa54-f6d67bc0dcb5 for instance with vm_state building and task_state spawning. [ 1386.776301] env[61868]: DEBUG oslo_vmware.api [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316175, 'name': Destroy_Task} progress is 100%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1386.982329] env[61868]: DEBUG nova.compute.utils [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1386.983775] env[61868]: DEBUG nova.compute.manager [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1386.983950] env[61868]: DEBUG nova.network.neutron [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1387.022649] env[61868]: DEBUG nova.policy [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4dbe667c15f14d13b6ef69bfe770ac70', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b3904d578a0e47d58eae75b1d4fbf41c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 1387.127967] env[61868]: DEBUG nova.network.neutron [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Successfully updated port: b1bf2dec-5f14-461f-aa54-f6d67bc0dcb5 {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1387.151777] env[61868]: DEBUG nova.compute.manager [req-0980a790-4ba5-4598-afff-8035896e7c66 req-c6977e5b-9003-4960-ae28-a28b78a613b4 service nova] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Received event network-changed-b1bf2dec-5f14-461f-aa54-f6d67bc0dcb5 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1387.152042] env[61868]: DEBUG nova.compute.manager [req-0980a790-4ba5-4598-afff-8035896e7c66 req-c6977e5b-9003-4960-ae28-a28b78a613b4 service nova] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Refreshing instance network info cache due to event network-changed-b1bf2dec-5f14-461f-aa54-f6d67bc0dcb5. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1387.152266] env[61868]: DEBUG oslo_concurrency.lockutils [req-0980a790-4ba5-4598-afff-8035896e7c66 req-c6977e5b-9003-4960-ae28-a28b78a613b4 service nova] Acquiring lock "refresh_cache-28a14ebd-6003-4e20-b556-2a3b939f15c8" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1387.152414] env[61868]: DEBUG oslo_concurrency.lockutils [req-0980a790-4ba5-4598-afff-8035896e7c66 req-c6977e5b-9003-4960-ae28-a28b78a613b4 service nova] Acquired lock "refresh_cache-28a14ebd-6003-4e20-b556-2a3b939f15c8" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1387.152573] env[61868]: DEBUG nova.network.neutron [req-0980a790-4ba5-4598-afff-8035896e7c66 req-c6977e5b-9003-4960-ae28-a28b78a613b4 service nova] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Refreshing network info cache for port b1bf2dec-5f14-461f-aa54-f6d67bc0dcb5 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1387.266663] env[61868]: DEBUG nova.network.neutron [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Successfully created port: dc8823b5-592e-40dc-81dd-5bad3cf56bce {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1387.277547] env[61868]: DEBUG oslo_vmware.api [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316175, 'name': Destroy_Task, 'duration_secs': 1.29899} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1387.277803] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Destroyed the VM [ 1387.278092] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Deleting Snapshot of the VM instance {{(pid=61868) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1387.278361] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-3b71be34-b400-417a-a9ed-94ca1583a0ca {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1387.284938] env[61868]: DEBUG oslo_vmware.api [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1387.284938] env[61868]: value = "task-1316176" [ 1387.284938] env[61868]: _type = "Task" [ 1387.284938] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1387.292680] env[61868]: DEBUG oslo_vmware.api [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316176, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1387.486572] env[61868]: DEBUG nova.compute.manager [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1387.631758] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquiring lock "refresh_cache-28a14ebd-6003-4e20-b556-2a3b939f15c8" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1387.683409] env[61868]: DEBUG nova.network.neutron [req-0980a790-4ba5-4598-afff-8035896e7c66 req-c6977e5b-9003-4960-ae28-a28b78a613b4 service nova] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1387.751483] env[61868]: DEBUG nova.network.neutron [req-0980a790-4ba5-4598-afff-8035896e7c66 req-c6977e5b-9003-4960-ae28-a28b78a613b4 service nova] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1387.794514] env[61868]: DEBUG oslo_vmware.api [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316176, 'name': RemoveSnapshot_Task, 'duration_secs': 0.34443} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1387.794786] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Deleted Snapshot of the VM instance {{(pid=61868) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1387.795071] env[61868]: DEBUG nova.compute.manager [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1387.795836] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0603452-45db-4016-8d0f-da856ae5f670 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1388.253933] env[61868]: DEBUG oslo_concurrency.lockutils [req-0980a790-4ba5-4598-afff-8035896e7c66 req-c6977e5b-9003-4960-ae28-a28b78a613b4 service nova] Releasing lock "refresh_cache-28a14ebd-6003-4e20-b556-2a3b939f15c8" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1388.254391] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquired lock "refresh_cache-28a14ebd-6003-4e20-b556-2a3b939f15c8" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1388.254559] env[61868]: DEBUG nova.network.neutron [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1388.307511] env[61868]: INFO nova.compute.manager [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Shelve offloading [ 1388.309194] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1388.309437] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c998520a-6c1f-422c-91c2-d7fb6fa5919a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1388.316472] env[61868]: DEBUG oslo_vmware.api [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1388.316472] env[61868]: value = "task-1316177" [ 1388.316472] env[61868]: _type = "Task" [ 1388.316472] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1388.324109] env[61868]: DEBUG oslo_vmware.api [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316177, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1388.495307] env[61868]: DEBUG nova.compute.manager [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1388.520688] env[61868]: DEBUG nova.virt.hardware [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1388.520950] env[61868]: DEBUG nova.virt.hardware [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1388.521125] env[61868]: DEBUG nova.virt.hardware [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1388.521313] env[61868]: DEBUG nova.virt.hardware [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1388.521463] env[61868]: DEBUG nova.virt.hardware [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1388.521615] env[61868]: DEBUG nova.virt.hardware [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1388.521822] env[61868]: DEBUG nova.virt.hardware [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1388.521984] env[61868]: DEBUG nova.virt.hardware [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1388.522164] env[61868]: DEBUG nova.virt.hardware [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1388.522330] env[61868]: DEBUG nova.virt.hardware [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1388.522503] env[61868]: DEBUG nova.virt.hardware [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1388.523646] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01080b68-7584-420f-ac33-d04d7dda1cb3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1388.531714] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b46e5397-4889-40f5-96bf-c7436907926e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1388.639491] env[61868]: DEBUG nova.compute.manager [req-46eb6df6-1082-4679-8bfd-55c5511ddb95 req-abe571f0-0362-40fc-9137-4f95049fbdcd service nova] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Received event network-vif-plugged-dc8823b5-592e-40dc-81dd-5bad3cf56bce {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1388.639699] env[61868]: DEBUG oslo_concurrency.lockutils [req-46eb6df6-1082-4679-8bfd-55c5511ddb95 req-abe571f0-0362-40fc-9137-4f95049fbdcd service nova] Acquiring lock "6ce00281-a201-445a-bb75-aac07fd481aa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1388.639889] env[61868]: DEBUG oslo_concurrency.lockutils [req-46eb6df6-1082-4679-8bfd-55c5511ddb95 req-abe571f0-0362-40fc-9137-4f95049fbdcd service nova] Lock "6ce00281-a201-445a-bb75-aac07fd481aa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1388.640083] env[61868]: DEBUG oslo_concurrency.lockutils [req-46eb6df6-1082-4679-8bfd-55c5511ddb95 req-abe571f0-0362-40fc-9137-4f95049fbdcd service nova] Lock "6ce00281-a201-445a-bb75-aac07fd481aa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1388.640257] env[61868]: DEBUG nova.compute.manager [req-46eb6df6-1082-4679-8bfd-55c5511ddb95 req-abe571f0-0362-40fc-9137-4f95049fbdcd service nova] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] No waiting events found dispatching network-vif-plugged-dc8823b5-592e-40dc-81dd-5bad3cf56bce {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1388.642910] env[61868]: WARNING nova.compute.manager [req-46eb6df6-1082-4679-8bfd-55c5511ddb95 req-abe571f0-0362-40fc-9137-4f95049fbdcd service nova] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Received unexpected event network-vif-plugged-dc8823b5-592e-40dc-81dd-5bad3cf56bce for instance with vm_state building and task_state spawning. [ 1388.721634] env[61868]: DEBUG nova.network.neutron [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Successfully updated port: dc8823b5-592e-40dc-81dd-5bad3cf56bce {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1388.783814] env[61868]: DEBUG nova.network.neutron [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1388.825929] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] VM already powered off {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1388.826159] env[61868]: DEBUG nova.compute.manager [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1388.826920] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07e4ac91-04c4-4598-a5e6-ac05942665a8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1388.832221] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "refresh_cache-b411db59-3058-4aa8-ac33-d3897c36c123" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1388.832391] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquired lock "refresh_cache-b411db59-3058-4aa8-ac33-d3897c36c123" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1388.832566] env[61868]: DEBUG nova.network.neutron [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1388.900626] env[61868]: DEBUG nova.network.neutron [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Updating instance_info_cache with network_info: [{"id": "b1bf2dec-5f14-461f-aa54-f6d67bc0dcb5", "address": "fa:16:3e:5c:1c:e5", "network": {"id": "5be7f7b7-0d32-48b1-979d-5a4a872e5b1a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1791355381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b3904d578a0e47d58eae75b1d4fbf41c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1bf2dec-5f", "ovs_interfaceid": "b1bf2dec-5f14-461f-aa54-f6d67bc0dcb5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1389.224286] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquiring lock "refresh_cache-6ce00281-a201-445a-bb75-aac07fd481aa" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1389.224418] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquired lock "refresh_cache-6ce00281-a201-445a-bb75-aac07fd481aa" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1389.224567] env[61868]: DEBUG nova.network.neutron [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1389.402831] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Releasing lock "refresh_cache-28a14ebd-6003-4e20-b556-2a3b939f15c8" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1389.403149] env[61868]: DEBUG nova.compute.manager [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Instance network_info: |[{"id": "b1bf2dec-5f14-461f-aa54-f6d67bc0dcb5", "address": "fa:16:3e:5c:1c:e5", "network": {"id": "5be7f7b7-0d32-48b1-979d-5a4a872e5b1a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1791355381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b3904d578a0e47d58eae75b1d4fbf41c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1bf2dec-5f", "ovs_interfaceid": "b1bf2dec-5f14-461f-aa54-f6d67bc0dcb5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1389.403567] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5c:1c:e5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a8b99a46-3e7f-4ef1-9e45-58e6cd17f210', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b1bf2dec-5f14-461f-aa54-f6d67bc0dcb5', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1389.410821] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Creating folder: Project (b3904d578a0e47d58eae75b1d4fbf41c). Parent ref: group-v281478. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1389.411103] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5867ce50-9dfb-48ac-981e-6755236435b4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1389.421555] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Created folder: Project (b3904d578a0e47d58eae75b1d4fbf41c) in parent group-v281478. [ 1389.421739] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Creating folder: Instances. Parent ref: group-v281652. {{(pid=61868) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1389.421960] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1cfb087d-50d3-4d38-a163-970a632803ce {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1389.430603] env[61868]: INFO nova.virt.vmwareapi.vm_util [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Created folder: Instances in parent group-v281652. [ 1389.430828] env[61868]: DEBUG oslo.service.loopingcall [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1389.431022] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1389.431211] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-01b9c8d9-0230-40a8-89df-ae74e279be7b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1389.451125] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1389.451125] env[61868]: value = "task-1316180" [ 1389.451125] env[61868]: _type = "Task" [ 1389.451125] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1389.458113] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1316180, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1389.539489] env[61868]: DEBUG nova.network.neutron [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Updating instance_info_cache with network_info: [{"id": "22435bb8-30a0-4fbb-b86c-a27082b82e4a", "address": "fa:16:3e:89:f5:a6", "network": {"id": "7e8cbecf-dfdd-43f2-8745-ae1d2d55d309", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-33935587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a7014141b24101ba7920bc83be4a9e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4db2961d-273d-4634-9d06-a94fa9d384fb", "external-id": "nsx-vlan-transportzone-572", "segmentation_id": 572, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22435bb8-30", "ovs_interfaceid": "22435bb8-30a0-4fbb-b86c-a27082b82e4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1389.755941] env[61868]: DEBUG nova.network.neutron [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1389.871514] env[61868]: DEBUG nova.network.neutron [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Updating instance_info_cache with network_info: [{"id": "dc8823b5-592e-40dc-81dd-5bad3cf56bce", "address": "fa:16:3e:c5:24:eb", "network": {"id": "5be7f7b7-0d32-48b1-979d-5a4a872e5b1a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1791355381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b3904d578a0e47d58eae75b1d4fbf41c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc8823b5-59", "ovs_interfaceid": "dc8823b5-592e-40dc-81dd-5bad3cf56bce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1389.960738] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1316180, 'name': CreateVM_Task, 'duration_secs': 0.270767} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1389.960876] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1389.961532] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1389.961705] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1389.962042] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1389.962318] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2b4d23d-df42-4548-bfd7-52a7ccbe8d47 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1389.966818] env[61868]: DEBUG oslo_vmware.api [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1389.966818] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]529a774b-d01c-8972-bfa9-3b632ae26696" [ 1389.966818] env[61868]: _type = "Task" [ 1389.966818] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1389.973934] env[61868]: DEBUG oslo_vmware.api [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]529a774b-d01c-8972-bfa9-3b632ae26696, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1390.041921] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Releasing lock "refresh_cache-b411db59-3058-4aa8-ac33-d3897c36c123" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1390.363304] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1390.364763] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6380f3cd-2118-4aff-91ec-2379757957c0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1390.371916] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1390.372176] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-92fffb22-ac78-4bbc-9e54-63bd89b7c0b4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1390.373832] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Releasing lock "refresh_cache-6ce00281-a201-445a-bb75-aac07fd481aa" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1390.374146] env[61868]: DEBUG nova.compute.manager [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Instance network_info: |[{"id": "dc8823b5-592e-40dc-81dd-5bad3cf56bce", "address": "fa:16:3e:c5:24:eb", "network": {"id": "5be7f7b7-0d32-48b1-979d-5a4a872e5b1a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1791355381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b3904d578a0e47d58eae75b1d4fbf41c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc8823b5-59", "ovs_interfaceid": "dc8823b5-592e-40dc-81dd-5bad3cf56bce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1390.374516] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c5:24:eb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a8b99a46-3e7f-4ef1-9e45-58e6cd17f210', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dc8823b5-592e-40dc-81dd-5bad3cf56bce', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1390.381864] env[61868]: DEBUG oslo.service.loopingcall [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1390.382433] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1390.382644] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-48aefcf3-4991-4867-ba86-fb9dee9d7071 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1390.401113] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1390.401113] env[61868]: value = "task-1316182" [ 1390.401113] env[61868]: _type = "Task" [ 1390.401113] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1390.408595] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1316182, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1390.450917] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1390.451204] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1390.451385] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Deleting the datastore file [datastore1] b411db59-3058-4aa8-ac33-d3897c36c123 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1390.451673] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4afcef3e-626b-47ec-8db0-993143d5ffe6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1390.457656] env[61868]: DEBUG oslo_vmware.api [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1390.457656] env[61868]: value = "task-1316183" [ 1390.457656] env[61868]: _type = "Task" [ 1390.457656] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1390.466799] env[61868]: DEBUG oslo_vmware.api [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316183, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1390.475993] env[61868]: DEBUG oslo_vmware.api [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]529a774b-d01c-8972-bfa9-3b632ae26696, 'name': SearchDatastore_Task, 'duration_secs': 0.009051} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1390.476301] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1390.476554] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1390.476820] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1390.476974] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1390.477171] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1390.477438] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e1f22728-b570-495d-b48b-57536500c25b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1390.492966] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1390.493134] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1390.493859] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a69606a4-7b8f-4932-9e92-38bb1effadd1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1390.499037] env[61868]: DEBUG oslo_vmware.api [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1390.499037] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52752bfd-3e17-2fb5-9051-94744bc970ab" [ 1390.499037] env[61868]: _type = "Task" [ 1390.499037] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1390.506855] env[61868]: DEBUG oslo_vmware.api [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52752bfd-3e17-2fb5-9051-94744bc970ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1390.669862] env[61868]: DEBUG nova.compute.manager [req-01d5e542-a3e9-46eb-9ed6-1ed736ce6f5e req-37487ecb-9bcd-4a9f-adc9-f74f8c2c4d6e service nova] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Received event network-changed-dc8823b5-592e-40dc-81dd-5bad3cf56bce {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1390.670133] env[61868]: DEBUG nova.compute.manager [req-01d5e542-a3e9-46eb-9ed6-1ed736ce6f5e req-37487ecb-9bcd-4a9f-adc9-f74f8c2c4d6e service nova] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Refreshing instance network info cache due to event network-changed-dc8823b5-592e-40dc-81dd-5bad3cf56bce. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1390.670354] env[61868]: DEBUG oslo_concurrency.lockutils [req-01d5e542-a3e9-46eb-9ed6-1ed736ce6f5e req-37487ecb-9bcd-4a9f-adc9-f74f8c2c4d6e service nova] Acquiring lock "refresh_cache-6ce00281-a201-445a-bb75-aac07fd481aa" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1390.670597] env[61868]: DEBUG oslo_concurrency.lockutils [req-01d5e542-a3e9-46eb-9ed6-1ed736ce6f5e req-37487ecb-9bcd-4a9f-adc9-f74f8c2c4d6e service nova] Acquired lock "refresh_cache-6ce00281-a201-445a-bb75-aac07fd481aa" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1390.670837] env[61868]: DEBUG nova.network.neutron [req-01d5e542-a3e9-46eb-9ed6-1ed736ce6f5e req-37487ecb-9bcd-4a9f-adc9-f74f8c2c4d6e service nova] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Refreshing network info cache for port dc8823b5-592e-40dc-81dd-5bad3cf56bce {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1390.910694] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1316182, 'name': CreateVM_Task, 'duration_secs': 0.288744} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1390.910871] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1390.911551] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1390.911727] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1390.912094] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1390.912352] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1f43577-6a9d-4f52-b725-c66c34de1c4a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1390.916351] env[61868]: DEBUG oslo_vmware.api [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1390.916351] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52f3de40-d5d4-17a9-9b87-fe4919692d56" [ 1390.916351] env[61868]: _type = "Task" [ 1390.916351] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1390.923742] env[61868]: DEBUG oslo_vmware.api [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52f3de40-d5d4-17a9-9b87-fe4919692d56, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1390.966520] env[61868]: DEBUG oslo_vmware.api [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316183, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150767} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1390.966773] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1390.966960] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1390.967157] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1390.990354] env[61868]: INFO nova.scheduler.client.report [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Deleted allocations for instance b411db59-3058-4aa8-ac33-d3897c36c123 [ 1391.007805] env[61868]: DEBUG oslo_vmware.api [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52752bfd-3e17-2fb5-9051-94744bc970ab, 'name': SearchDatastore_Task, 'duration_secs': 0.008795} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1391.008649] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb5fb8b0-e321-42fc-8725-04ccde2e9693 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1391.013646] env[61868]: DEBUG oslo_vmware.api [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1391.013646] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]526b992c-7549-bed8-c50f-bcd4b2fcacc8" [ 1391.013646] env[61868]: _type = "Task" [ 1391.013646] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1391.021121] env[61868]: DEBUG oslo_vmware.api [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]526b992c-7549-bed8-c50f-bcd4b2fcacc8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1391.354551] env[61868]: DEBUG nova.network.neutron [req-01d5e542-a3e9-46eb-9ed6-1ed736ce6f5e req-37487ecb-9bcd-4a9f-adc9-f74f8c2c4d6e service nova] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Updated VIF entry in instance network info cache for port dc8823b5-592e-40dc-81dd-5bad3cf56bce. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1391.354986] env[61868]: DEBUG nova.network.neutron [req-01d5e542-a3e9-46eb-9ed6-1ed736ce6f5e req-37487ecb-9bcd-4a9f-adc9-f74f8c2c4d6e service nova] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Updating instance_info_cache with network_info: [{"id": "dc8823b5-592e-40dc-81dd-5bad3cf56bce", "address": "fa:16:3e:c5:24:eb", "network": {"id": "5be7f7b7-0d32-48b1-979d-5a4a872e5b1a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1791355381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b3904d578a0e47d58eae75b1d4fbf41c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc8823b5-59", "ovs_interfaceid": "dc8823b5-592e-40dc-81dd-5bad3cf56bce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1391.426828] env[61868]: DEBUG oslo_vmware.api [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52f3de40-d5d4-17a9-9b87-fe4919692d56, 'name': SearchDatastore_Task, 'duration_secs': 0.009685} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1391.427105] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1391.427346] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1391.427590] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1391.495529] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1391.495900] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1391.496226] env[61868]: DEBUG nova.objects.instance [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lazy-loading 'resources' on Instance uuid b411db59-3058-4aa8-ac33-d3897c36c123 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1391.524059] env[61868]: DEBUG oslo_vmware.api [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]526b992c-7549-bed8-c50f-bcd4b2fcacc8, 'name': SearchDatastore_Task, 'duration_secs': 0.008696} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1391.524059] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1391.524237] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 28a14ebd-6003-4e20-b556-2a3b939f15c8/28a14ebd-6003-4e20-b556-2a3b939f15c8.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1391.524424] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1391.524810] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1391.524810] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6f5ac916-59c0-49ad-9b95-f42618d528bb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1391.527750] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1c53d58f-9aab-42ca-9e18-83bc98541c2a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1391.533363] env[61868]: DEBUG oslo_vmware.api [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1391.533363] env[61868]: value = "task-1316184" [ 1391.533363] env[61868]: _type = "Task" [ 1391.533363] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1391.537080] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1391.537261] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1391.538179] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a318ab74-a914-4575-b9f8-5338075b5112 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1391.542874] env[61868]: DEBUG oslo_vmware.api [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316184, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1391.545608] env[61868]: DEBUG oslo_vmware.api [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1391.545608] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5211c710-54b9-d9b9-2c6d-e49556b2b3e1" [ 1391.545608] env[61868]: _type = "Task" [ 1391.545608] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1391.552620] env[61868]: DEBUG oslo_vmware.api [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5211c710-54b9-d9b9-2c6d-e49556b2b3e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1391.857509] env[61868]: DEBUG oslo_concurrency.lockutils [req-01d5e542-a3e9-46eb-9ed6-1ed736ce6f5e req-37487ecb-9bcd-4a9f-adc9-f74f8c2c4d6e service nova] Releasing lock "refresh_cache-6ce00281-a201-445a-bb75-aac07fd481aa" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1391.857835] env[61868]: DEBUG nova.compute.manager [req-01d5e542-a3e9-46eb-9ed6-1ed736ce6f5e req-37487ecb-9bcd-4a9f-adc9-f74f8c2c4d6e service nova] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Received event network-vif-unplugged-22435bb8-30a0-4fbb-b86c-a27082b82e4a {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1391.858021] env[61868]: DEBUG oslo_concurrency.lockutils [req-01d5e542-a3e9-46eb-9ed6-1ed736ce6f5e req-37487ecb-9bcd-4a9f-adc9-f74f8c2c4d6e service nova] Acquiring lock "b411db59-3058-4aa8-ac33-d3897c36c123-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1391.858254] env[61868]: DEBUG oslo_concurrency.lockutils [req-01d5e542-a3e9-46eb-9ed6-1ed736ce6f5e req-37487ecb-9bcd-4a9f-adc9-f74f8c2c4d6e service nova] Lock "b411db59-3058-4aa8-ac33-d3897c36c123-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1391.858417] env[61868]: DEBUG oslo_concurrency.lockutils [req-01d5e542-a3e9-46eb-9ed6-1ed736ce6f5e req-37487ecb-9bcd-4a9f-adc9-f74f8c2c4d6e service nova] Lock "b411db59-3058-4aa8-ac33-d3897c36c123-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1391.858586] env[61868]: DEBUG nova.compute.manager [req-01d5e542-a3e9-46eb-9ed6-1ed736ce6f5e req-37487ecb-9bcd-4a9f-adc9-f74f8c2c4d6e service nova] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] No waiting events found dispatching network-vif-unplugged-22435bb8-30a0-4fbb-b86c-a27082b82e4a {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1391.858771] env[61868]: WARNING nova.compute.manager [req-01d5e542-a3e9-46eb-9ed6-1ed736ce6f5e req-37487ecb-9bcd-4a9f-adc9-f74f8c2c4d6e service nova] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Received unexpected event network-vif-unplugged-22435bb8-30a0-4fbb-b86c-a27082b82e4a for instance with vm_state shelved and task_state shelving_offloading. [ 1391.858937] env[61868]: DEBUG nova.compute.manager [req-01d5e542-a3e9-46eb-9ed6-1ed736ce6f5e req-37487ecb-9bcd-4a9f-adc9-f74f8c2c4d6e service nova] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Received event network-changed-22435bb8-30a0-4fbb-b86c-a27082b82e4a {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1391.859122] env[61868]: DEBUG nova.compute.manager [req-01d5e542-a3e9-46eb-9ed6-1ed736ce6f5e req-37487ecb-9bcd-4a9f-adc9-f74f8c2c4d6e service nova] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Refreshing instance network info cache due to event network-changed-22435bb8-30a0-4fbb-b86c-a27082b82e4a. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1391.859324] env[61868]: DEBUG oslo_concurrency.lockutils [req-01d5e542-a3e9-46eb-9ed6-1ed736ce6f5e req-37487ecb-9bcd-4a9f-adc9-f74f8c2c4d6e service nova] Acquiring lock "refresh_cache-b411db59-3058-4aa8-ac33-d3897c36c123" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1391.859477] env[61868]: DEBUG oslo_concurrency.lockutils [req-01d5e542-a3e9-46eb-9ed6-1ed736ce6f5e req-37487ecb-9bcd-4a9f-adc9-f74f8c2c4d6e service nova] Acquired lock "refresh_cache-b411db59-3058-4aa8-ac33-d3897c36c123" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1391.859618] env[61868]: DEBUG nova.network.neutron [req-01d5e542-a3e9-46eb-9ed6-1ed736ce6f5e req-37487ecb-9bcd-4a9f-adc9-f74f8c2c4d6e service nova] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Refreshing network info cache for port 22435bb8-30a0-4fbb-b86c-a27082b82e4a {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1391.998609] env[61868]: DEBUG nova.objects.instance [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lazy-loading 'numa_topology' on Instance uuid b411db59-3058-4aa8-ac33-d3897c36c123 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1392.043147] env[61868]: DEBUG oslo_vmware.api [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316184, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.451354} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1392.043401] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 28a14ebd-6003-4e20-b556-2a3b939f15c8/28a14ebd-6003-4e20-b556-2a3b939f15c8.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1392.043614] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1392.043859] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e29aba77-98d0-4997-909f-40405db3aecc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1392.049989] env[61868]: DEBUG oslo_vmware.api [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1392.049989] env[61868]: value = "task-1316185" [ 1392.049989] env[61868]: _type = "Task" [ 1392.049989] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1392.055812] env[61868]: DEBUG oslo_vmware.api [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5211c710-54b9-d9b9-2c6d-e49556b2b3e1, 'name': SearchDatastore_Task, 'duration_secs': 0.008144} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1392.056809] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-90316b8f-5e60-4dbe-84ce-94cc18dd6c5b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1392.061672] env[61868]: DEBUG oslo_vmware.api [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316185, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1392.064441] env[61868]: DEBUG oslo_vmware.api [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1392.064441] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]527b5b02-e2fd-1a3c-dc63-8ca579d34249" [ 1392.064441] env[61868]: _type = "Task" [ 1392.064441] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1392.071300] env[61868]: DEBUG oslo_vmware.api [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]527b5b02-e2fd-1a3c-dc63-8ca579d34249, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1392.500957] env[61868]: DEBUG nova.objects.base [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=61868) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1392.546283] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fce9e42b-782d-47f6-8362-7315599bfd99 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1392.555980] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83e7a759-0700-4df9-bd09-a7cc7398cd60 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1392.563328] env[61868]: DEBUG oslo_vmware.api [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316185, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058105} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1392.586916] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1392.588135] env[61868]: DEBUG nova.network.neutron [req-01d5e542-a3e9-46eb-9ed6-1ed736ce6f5e req-37487ecb-9bcd-4a9f-adc9-f74f8c2c4d6e service nova] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Updated VIF entry in instance network info cache for port 22435bb8-30a0-4fbb-b86c-a27082b82e4a. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1392.588454] env[61868]: DEBUG nova.network.neutron [req-01d5e542-a3e9-46eb-9ed6-1ed736ce6f5e req-37487ecb-9bcd-4a9f-adc9-f74f8c2c4d6e service nova] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Updating instance_info_cache with network_info: [{"id": "22435bb8-30a0-4fbb-b86c-a27082b82e4a", "address": "fa:16:3e:89:f5:a6", "network": {"id": "7e8cbecf-dfdd-43f2-8745-ae1d2d55d309", "bridge": null, "label": "tempest-ServerActionsTestOtherB-33935587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a7014141b24101ba7920bc83be4a9e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap22435bb8-30", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1392.589950] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07a34c1d-fecb-4b8c-a1b9-4f0d1584c8ba {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1392.595659] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afe238d7-f3f5-4f64-85eb-ac19cc72f941 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1392.617015] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Reconfiguring VM instance instance-00000070 to attach disk [datastore2] 28a14ebd-6003-4e20-b556-2a3b939f15c8/28a14ebd-6003-4e20-b556-2a3b939f15c8.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1392.622367] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ccb628b7-4913-46a8-87a9-f8b436bcbeff {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1392.635644] env[61868]: DEBUG oslo_vmware.api [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]527b5b02-e2fd-1a3c-dc63-8ca579d34249, 'name': SearchDatastore_Task, 'duration_secs': 0.008109} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1392.636184] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1392.636436] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 6ce00281-a201-445a-bb75-aac07fd481aa/6ce00281-a201-445a-bb75-aac07fd481aa.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1392.637596] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-674418c9-c217-43fa-9d3f-e50614cfb693 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1392.641563] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bbe704c1-0c4c-4363-ba74-e4c4a626f142 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1392.645977] env[61868]: DEBUG oslo_vmware.api [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1392.645977] env[61868]: value = "task-1316186" [ 1392.645977] env[61868]: _type = "Task" [ 1392.645977] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1392.654332] env[61868]: DEBUG nova.compute.provider_tree [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1392.659392] env[61868]: DEBUG oslo_vmware.api [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1392.659392] env[61868]: value = "task-1316187" [ 1392.659392] env[61868]: _type = "Task" [ 1392.659392] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1392.666343] env[61868]: DEBUG oslo_vmware.api [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316186, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1392.670842] env[61868]: DEBUG oslo_vmware.api [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316187, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1392.801482] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "b411db59-3058-4aa8-ac33-d3897c36c123" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1393.099952] env[61868]: DEBUG oslo_concurrency.lockutils [req-01d5e542-a3e9-46eb-9ed6-1ed736ce6f5e req-37487ecb-9bcd-4a9f-adc9-f74f8c2c4d6e service nova] Releasing lock "refresh_cache-b411db59-3058-4aa8-ac33-d3897c36c123" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1393.158987] env[61868]: DEBUG nova.scheduler.client.report [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1393.170578] env[61868]: DEBUG oslo_vmware.api [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316186, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1393.175428] env[61868]: DEBUG oslo_vmware.api [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316187, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.476226} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1393.176283] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 6ce00281-a201-445a-bb75-aac07fd481aa/6ce00281-a201-445a-bb75-aac07fd481aa.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1393.176510] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1393.176768] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-25b9e6ce-10c7-4d98-bf45-b0d60a6a9db6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1393.182265] env[61868]: DEBUG oslo_vmware.api [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1393.182265] env[61868]: value = "task-1316188" [ 1393.182265] env[61868]: _type = "Task" [ 1393.182265] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1393.189548] env[61868]: DEBUG oslo_vmware.api [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316188, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1393.663943] env[61868]: DEBUG oslo_vmware.api [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316186, 'name': ReconfigVM_Task, 'duration_secs': 0.539578} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1393.664227] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Reconfigured VM instance instance-00000070 to attach disk [datastore2] 28a14ebd-6003-4e20-b556-2a3b939f15c8/28a14ebd-6003-4e20-b556-2a3b939f15c8.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1393.664853] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6d7535a5-7b6c-4f60-ad3e-49a39c4bed5f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1393.666712] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.171s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1393.673417] env[61868]: DEBUG oslo_vmware.api [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1393.673417] env[61868]: value = "task-1316189" [ 1393.673417] env[61868]: _type = "Task" [ 1393.673417] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1393.680775] env[61868]: DEBUG oslo_vmware.api [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316189, 'name': Rename_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1393.690220] env[61868]: DEBUG oslo_vmware.api [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316188, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062089} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1393.690460] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1393.691235] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-720f8e89-33b8-48c5-8028-e7eac470d1f8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1393.714756] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Reconfiguring VM instance instance-00000071 to attach disk [datastore2] 6ce00281-a201-445a-bb75-aac07fd481aa/6ce00281-a201-445a-bb75-aac07fd481aa.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1393.715609] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6808db23-424b-49d8-9a78-e2e1c067e561 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1393.734163] env[61868]: DEBUG oslo_vmware.api [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1393.734163] env[61868]: value = "task-1316190" [ 1393.734163] env[61868]: _type = "Task" [ 1393.734163] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1393.741783] env[61868]: DEBUG oslo_vmware.api [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316190, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1394.178385] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a6bd95a8-cfa9-4d3c-ac2a-992e50b23507 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "b411db59-3058-4aa8-ac33-d3897c36c123" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 20.292s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1394.179612] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "b411db59-3058-4aa8-ac33-d3897c36c123" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.378s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1394.179923] env[61868]: INFO nova.compute.manager [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Unshelving [ 1394.186345] env[61868]: DEBUG oslo_vmware.api [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316189, 'name': Rename_Task, 'duration_secs': 0.130691} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1394.186630] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1394.187678] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-949bc75a-8682-4255-9033-58bce1da2ac8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1394.193625] env[61868]: DEBUG oslo_vmware.api [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1394.193625] env[61868]: value = "task-1316191" [ 1394.193625] env[61868]: _type = "Task" [ 1394.193625] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1394.201217] env[61868]: DEBUG oslo_vmware.api [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316191, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1394.243583] env[61868]: DEBUG oslo_vmware.api [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316190, 'name': ReconfigVM_Task, 'duration_secs': 0.251219} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1394.243902] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Reconfigured VM instance instance-00000071 to attach disk [datastore2] 6ce00281-a201-445a-bb75-aac07fd481aa/6ce00281-a201-445a-bb75-aac07fd481aa.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1394.244487] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-15badef1-7790-4605-b3bb-a635a7ed8cf2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1394.252121] env[61868]: DEBUG oslo_vmware.api [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1394.252121] env[61868]: value = "task-1316192" [ 1394.252121] env[61868]: _type = "Task" [ 1394.252121] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1394.260044] env[61868]: DEBUG oslo_vmware.api [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316192, 'name': Rename_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1394.703453] env[61868]: DEBUG oslo_vmware.api [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316191, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1394.762804] env[61868]: DEBUG oslo_vmware.api [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316192, 'name': Rename_Task, 'duration_secs': 0.135374} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1394.764030] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1394.764030] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8938b2d3-3c06-42c1-b4be-8aeb3cd5faf0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1394.769538] env[61868]: DEBUG oslo_vmware.api [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1394.769538] env[61868]: value = "task-1316193" [ 1394.769538] env[61868]: _type = "Task" [ 1394.769538] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1394.777378] env[61868]: DEBUG oslo_vmware.api [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316193, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1395.204396] env[61868]: DEBUG oslo_vmware.api [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316191, 'name': PowerOnVM_Task, 'duration_secs': 0.638229} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1395.205733] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1395.205975] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1395.206207] env[61868]: DEBUG nova.objects.instance [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lazy-loading 'pci_requests' on Instance uuid b411db59-3058-4aa8-ac33-d3897c36c123 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1395.207283] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1395.207472] env[61868]: INFO nova.compute.manager [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Took 8.84 seconds to spawn the instance on the hypervisor. [ 1395.207683] env[61868]: DEBUG nova.compute.manager [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1395.208694] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-624b0fca-8969-467a-9dbd-892d283591e8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1395.279297] env[61868]: DEBUG oslo_vmware.api [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316193, 'name': PowerOnVM_Task, 'duration_secs': 0.49029} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1395.279675] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1395.279876] env[61868]: INFO nova.compute.manager [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Took 6.78 seconds to spawn the instance on the hypervisor. [ 1395.280075] env[61868]: DEBUG nova.compute.manager [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1395.280831] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de81d57b-d49f-439c-9ac8-92a6949465fa {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1395.710547] env[61868]: DEBUG nova.objects.instance [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lazy-loading 'numa_topology' on Instance uuid b411db59-3058-4aa8-ac33-d3897c36c123 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1395.724241] env[61868]: INFO nova.compute.manager [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Took 13.54 seconds to build instance. [ 1395.795753] env[61868]: INFO nova.compute.manager [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Took 13.28 seconds to build instance. [ 1396.075455] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1396.213612] env[61868]: INFO nova.compute.claims [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1396.226448] env[61868]: DEBUG oslo_concurrency.lockutils [None req-0785f897-fa98-49a2-9da7-348a1bfb61e4 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lock "28a14ebd-6003-4e20-b556-2a3b939f15c8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.047s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1396.297837] env[61868]: DEBUG oslo_concurrency.lockutils [None req-03dc8754-6fd1-4d59-8fca-d82fb3371b99 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lock "6ce00281-a201-445a-bb75-aac07fd481aa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.790s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1396.920135] env[61868]: INFO nova.compute.manager [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Rescuing [ 1396.920428] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquiring lock "refresh_cache-6ce00281-a201-445a-bb75-aac07fd481aa" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1396.920590] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquired lock "refresh_cache-6ce00281-a201-445a-bb75-aac07fd481aa" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1396.920764] env[61868]: DEBUG nova.network.neutron [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1397.074871] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1397.075090] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61868) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1397.270706] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca5e606d-6930-4ff6-a85c-4b3030c6a206 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1397.279810] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eb9b7ee-b458-4f87-a312-4d1044aeb5a2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1397.309087] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d725f59-93de-4ada-a7a3-c1624471a6e6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1397.316258] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-923b639a-a388-44ec-b246-ad02e049f07e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1397.329246] env[61868]: DEBUG nova.compute.provider_tree [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1397.609405] env[61868]: DEBUG nova.network.neutron [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Updating instance_info_cache with network_info: [{"id": "dc8823b5-592e-40dc-81dd-5bad3cf56bce", "address": "fa:16:3e:c5:24:eb", "network": {"id": "5be7f7b7-0d32-48b1-979d-5a4a872e5b1a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1791355381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b3904d578a0e47d58eae75b1d4fbf41c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc8823b5-59", "ovs_interfaceid": "dc8823b5-592e-40dc-81dd-5bad3cf56bce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1397.832643] env[61868]: DEBUG nova.scheduler.client.report [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1398.075347] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager.update_available_resource {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1398.113239] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Releasing lock "refresh_cache-6ce00281-a201-445a-bb75-aac07fd481aa" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1398.337771] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.132s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1398.365026] env[61868]: INFO nova.network.neutron [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Updating port 22435bb8-30a0-4fbb-b86c-a27082b82e4a with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1398.578200] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1398.578442] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1398.578636] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1398.578922] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61868) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1398.579701] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dc33878-5a80-4a47-a1bb-2fa7b1b5771e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1398.588880] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f638383a-38d2-46c0-adb2-016255ebf5b6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1398.602605] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87cbdf6f-e872-4699-9b47-9748ce7c47b4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1398.608747] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f0bfc29-d2d9-4647-ba96-b111f2a8626f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1398.636342] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180934MB free_disk=134GB free_vcpus=48 pci_devices=None {{(pid=61868) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1398.636508] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1398.636679] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1398.642937] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1398.643184] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-812cd426-8797-4d80-a360-f20a3b570aad {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1398.648996] env[61868]: DEBUG oslo_vmware.api [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1398.648996] env[61868]: value = "task-1316194" [ 1398.648996] env[61868]: _type = "Task" [ 1398.648996] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1398.656400] env[61868]: DEBUG oslo_vmware.api [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316194, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1399.159466] env[61868]: DEBUG oslo_vmware.api [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316194, 'name': PowerOffVM_Task, 'duration_secs': 0.194732} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1399.159719] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1399.160471] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec40d6af-64d7-4a72-b27b-4afbe1a3b8fa {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1399.179096] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02900902-c123-4501-82cc-acc3534dd089 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1399.204179] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1399.204436] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3f1038de-07d4-4f60-a4aa-990fafbb663e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1399.210799] env[61868]: DEBUG oslo_vmware.api [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1399.210799] env[61868]: value = "task-1316195" [ 1399.210799] env[61868]: _type = "Task" [ 1399.210799] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1399.217848] env[61868]: DEBUG oslo_vmware.api [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316195, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1399.661692] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 28a14ebd-6003-4e20-b556-2a3b939f15c8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1399.661970] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 6ce00281-a201-445a-bb75-aac07fd481aa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1399.661970] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance b411db59-3058-4aa8-ac33-d3897c36c123 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1399.662125] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=61868) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1399.662277] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=61868) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1399.709013] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3888c0f8-c43d-4394-9a61-8e0322ead518 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1399.721721] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-905fd4fb-f82d-4e89-8297-41eff8f3956e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1399.724619] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] VM already powered off {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1399.724822] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1399.725073] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1399.725229] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1399.725408] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1399.726113] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e8e2e55c-ba87-4ec6-b334-148242d74b9e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1399.751649] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93d7fce8-bf85-47d8-8a1e-1fc502e51af9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1399.754007] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1399.754198] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1399.754848] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39df20ef-e64b-48c2-a008-44f7beb395ef {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1399.762016] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f5c46f8-7bb5-4358-a5e3-a6314ff38db4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1399.765460] env[61868]: DEBUG oslo_vmware.api [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1399.765460] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]526d2e32-8492-f5b7-7d37-0d3f45eea430" [ 1399.765460] env[61868]: _type = "Task" [ 1399.765460] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1399.775527] env[61868]: DEBUG nova.compute.provider_tree [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1399.781380] env[61868]: DEBUG oslo_vmware.api [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]526d2e32-8492-f5b7-7d37-0d3f45eea430, 'name': SearchDatastore_Task, 'duration_secs': 0.008434} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1399.782074] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ee93dee-cfce-494d-ab8e-daba663b71f6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1399.787048] env[61868]: DEBUG oslo_vmware.api [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1399.787048] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]528c3db2-69e0-9a31-76ce-8cc03414cdc4" [ 1399.787048] env[61868]: _type = "Task" [ 1399.787048] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1399.795275] env[61868]: DEBUG oslo_vmware.api [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]528c3db2-69e0-9a31-76ce-8cc03414cdc4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1399.921723] env[61868]: DEBUG nova.compute.manager [req-5f6f3c3c-a494-4def-922b-d411d04f482a req-20a1c87c-1cb6-4354-b41c-7106ec6e6105 service nova] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Received event network-vif-plugged-22435bb8-30a0-4fbb-b86c-a27082b82e4a {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1399.921957] env[61868]: DEBUG oslo_concurrency.lockutils [req-5f6f3c3c-a494-4def-922b-d411d04f482a req-20a1c87c-1cb6-4354-b41c-7106ec6e6105 service nova] Acquiring lock "b411db59-3058-4aa8-ac33-d3897c36c123-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1399.922187] env[61868]: DEBUG oslo_concurrency.lockutils [req-5f6f3c3c-a494-4def-922b-d411d04f482a req-20a1c87c-1cb6-4354-b41c-7106ec6e6105 service nova] Lock "b411db59-3058-4aa8-ac33-d3897c36c123-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1399.922357] env[61868]: DEBUG oslo_concurrency.lockutils [req-5f6f3c3c-a494-4def-922b-d411d04f482a req-20a1c87c-1cb6-4354-b41c-7106ec6e6105 service nova] Lock "b411db59-3058-4aa8-ac33-d3897c36c123-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1399.922528] env[61868]: DEBUG nova.compute.manager [req-5f6f3c3c-a494-4def-922b-d411d04f482a req-20a1c87c-1cb6-4354-b41c-7106ec6e6105 service nova] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] No waiting events found dispatching network-vif-plugged-22435bb8-30a0-4fbb-b86c-a27082b82e4a {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1399.922798] env[61868]: WARNING nova.compute.manager [req-5f6f3c3c-a494-4def-922b-d411d04f482a req-20a1c87c-1cb6-4354-b41c-7106ec6e6105 service nova] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Received unexpected event network-vif-plugged-22435bb8-30a0-4fbb-b86c-a27082b82e4a for instance with vm_state shelved_offloaded and task_state spawning. [ 1400.013324] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "refresh_cache-b411db59-3058-4aa8-ac33-d3897c36c123" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1400.013582] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquired lock "refresh_cache-b411db59-3058-4aa8-ac33-d3897c36c123" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1400.013769] env[61868]: DEBUG nova.network.neutron [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1400.278777] env[61868]: DEBUG nova.scheduler.client.report [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1400.297404] env[61868]: DEBUG oslo_vmware.api [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]528c3db2-69e0-9a31-76ce-8cc03414cdc4, 'name': SearchDatastore_Task, 'duration_secs': 0.00835} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1400.297685] env[61868]: DEBUG oslo_concurrency.lockutils [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1400.297940] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 6ce00281-a201-445a-bb75-aac07fd481aa/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d-rescue.vmdk. {{(pid=61868) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1400.298205] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c1197336-4e3a-43bb-a2be-aa64c60fd56c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1400.304363] env[61868]: DEBUG oslo_vmware.api [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1400.304363] env[61868]: value = "task-1316196" [ 1400.304363] env[61868]: _type = "Task" [ 1400.304363] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1400.311686] env[61868]: DEBUG oslo_vmware.api [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316196, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1400.758594] env[61868]: DEBUG nova.network.neutron [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Updating instance_info_cache with network_info: [{"id": "22435bb8-30a0-4fbb-b86c-a27082b82e4a", "address": "fa:16:3e:89:f5:a6", "network": {"id": "7e8cbecf-dfdd-43f2-8745-ae1d2d55d309", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-33935587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a7014141b24101ba7920bc83be4a9e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4db2961d-273d-4634-9d06-a94fa9d384fb", "external-id": "nsx-vlan-transportzone-572", "segmentation_id": 572, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22435bb8-30", "ovs_interfaceid": "22435bb8-30a0-4fbb-b86c-a27082b82e4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1400.782900] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61868) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1400.783125] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.146s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1400.783332] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1400.783471] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Cleaning up deleted instances {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1400.814531] env[61868]: DEBUG oslo_vmware.api [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316196, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.446333} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1400.815729] env[61868]: INFO nova.virt.vmwareapi.ds_util [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 6ce00281-a201-445a-bb75-aac07fd481aa/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d-rescue.vmdk. [ 1400.816487] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6aac05cb-f0f4-439c-b5b4-f159e5679e1b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1400.841107] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Reconfiguring VM instance instance-00000071 to attach disk [datastore2] 6ce00281-a201-445a-bb75-aac07fd481aa/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d-rescue.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1400.841341] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bcff3b3d-af92-40ea-b254-319bfec870cf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1400.858194] env[61868]: DEBUG oslo_vmware.api [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1400.858194] env[61868]: value = "task-1316197" [ 1400.858194] env[61868]: _type = "Task" [ 1400.858194] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1400.865181] env[61868]: DEBUG oslo_vmware.api [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316197, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1401.261668] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Releasing lock "refresh_cache-b411db59-3058-4aa8-ac33-d3897c36c123" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1401.287067] env[61868]: DEBUG nova.virt.hardware [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='db10a32ec89a64afb956e96b9f4adc12',container_format='bare',created_at=2024-10-18T17:33:02Z,direct_url=,disk_format='vmdk',id=6308358b-8a38-4242-ad10-cd94d2d8a7fb,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-1207533121-shelved',owner='28a7014141b24101ba7920bc83be4a9e',properties=ImageMetaProps,protected=,size=31668224,status='active',tags=,updated_at=2024-10-18T17:33:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1401.287334] env[61868]: DEBUG nova.virt.hardware [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1401.287496] env[61868]: DEBUG nova.virt.hardware [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1401.287718] env[61868]: DEBUG nova.virt.hardware [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1401.287876] env[61868]: DEBUG nova.virt.hardware [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1401.288042] env[61868]: DEBUG nova.virt.hardware [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1401.288259] env[61868]: DEBUG nova.virt.hardware [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1401.288422] env[61868]: DEBUG nova.virt.hardware [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1401.288607] env[61868]: DEBUG nova.virt.hardware [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1401.288797] env[61868]: DEBUG nova.virt.hardware [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1401.288973] env[61868]: DEBUG nova.virt.hardware [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1401.291519] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] There are 9 instances to clean {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1401.291677] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 77e19bcd-5963-4c48-bcf2-da5c92b142ff] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1401.293883] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5d16e98-c4c0-4a6d-acbd-fe35baa709e8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1401.302169] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb734c32-fd2a-45de-90bd-5243ec1ab871 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1401.316801] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:89:f5:a6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4db2961d-273d-4634-9d06-a94fa9d384fb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '22435bb8-30a0-4fbb-b86c-a27082b82e4a', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1401.323917] env[61868]: DEBUG oslo.service.loopingcall [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1401.324767] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1401.324985] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-aec53e65-2517-42f8-8542-b11f2f273a80 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1401.343616] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1401.343616] env[61868]: value = "task-1316198" [ 1401.343616] env[61868]: _type = "Task" [ 1401.343616] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1401.350804] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1316198, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1401.366440] env[61868]: DEBUG oslo_vmware.api [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316197, 'name': ReconfigVM_Task, 'duration_secs': 0.258575} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1401.366701] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Reconfigured VM instance instance-00000071 to attach disk [datastore2] 6ce00281-a201-445a-bb75-aac07fd481aa/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d-rescue.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1401.367471] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-867b0166-1603-485b-b220-509f72d7c4b3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1401.392153] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-33e0b5a1-8ca1-48cd-8b88-8bdf8c920f8d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1401.406778] env[61868]: DEBUG oslo_vmware.api [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1401.406778] env[61868]: value = "task-1316199" [ 1401.406778] env[61868]: _type = "Task" [ 1401.406778] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1401.414561] env[61868]: DEBUG oslo_vmware.api [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316199, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1401.798194] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 32abbff4-5d9f-4c4c-8270-e3e5d383825e] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1401.853710] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1316198, 'name': CreateVM_Task, 'duration_secs': 0.331486} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1401.853892] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1401.854575] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6308358b-8a38-4242-ad10-cd94d2d8a7fb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1401.854745] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6308358b-8a38-4242-ad10-cd94d2d8a7fb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1401.855147] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6308358b-8a38-4242-ad10-cd94d2d8a7fb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1401.855398] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ccdb85d2-55e2-4ac2-bec6-6159824c6e41 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1401.859752] env[61868]: DEBUG oslo_vmware.api [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1401.859752] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52eb9b69-3b36-e89c-5fd6-ae54c7af06e4" [ 1401.859752] env[61868]: _type = "Task" [ 1401.859752] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1401.866881] env[61868]: DEBUG oslo_vmware.api [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52eb9b69-3b36-e89c-5fd6-ae54c7af06e4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1401.915552] env[61868]: DEBUG oslo_vmware.api [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316199, 'name': ReconfigVM_Task, 'duration_secs': 0.426414} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1401.915825] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1401.916086] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0002b685-05b1-4c9f-938e-d8445747952b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1401.921561] env[61868]: DEBUG oslo_vmware.api [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1401.921561] env[61868]: value = "task-1316200" [ 1401.921561] env[61868]: _type = "Task" [ 1401.921561] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1401.928654] env[61868]: DEBUG oslo_vmware.api [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316200, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1401.947139] env[61868]: DEBUG nova.compute.manager [req-fed27476-ba7c-4bfa-b2f2-42760a9b6f48 req-ebf9864b-0c85-4e0c-8df1-9167977037f8 service nova] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Received event network-changed-22435bb8-30a0-4fbb-b86c-a27082b82e4a {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1401.947290] env[61868]: DEBUG nova.compute.manager [req-fed27476-ba7c-4bfa-b2f2-42760a9b6f48 req-ebf9864b-0c85-4e0c-8df1-9167977037f8 service nova] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Refreshing instance network info cache due to event network-changed-22435bb8-30a0-4fbb-b86c-a27082b82e4a. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1401.947515] env[61868]: DEBUG oslo_concurrency.lockutils [req-fed27476-ba7c-4bfa-b2f2-42760a9b6f48 req-ebf9864b-0c85-4e0c-8df1-9167977037f8 service nova] Acquiring lock "refresh_cache-b411db59-3058-4aa8-ac33-d3897c36c123" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1401.947719] env[61868]: DEBUG oslo_concurrency.lockutils [req-fed27476-ba7c-4bfa-b2f2-42760a9b6f48 req-ebf9864b-0c85-4e0c-8df1-9167977037f8 service nova] Acquired lock "refresh_cache-b411db59-3058-4aa8-ac33-d3897c36c123" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1401.947893] env[61868]: DEBUG nova.network.neutron [req-fed27476-ba7c-4bfa-b2f2-42760a9b6f48 req-ebf9864b-0c85-4e0c-8df1-9167977037f8 service nova] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Refreshing network info cache for port 22435bb8-30a0-4fbb-b86c-a27082b82e4a {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1402.301658] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: e18dfd04-5a37-44d1-930b-52b50362755b] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1402.370168] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6308358b-8a38-4242-ad10-cd94d2d8a7fb" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1402.370394] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Processing image 6308358b-8a38-4242-ad10-cd94d2d8a7fb {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1402.370614] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6308358b-8a38-4242-ad10-cd94d2d8a7fb/6308358b-8a38-4242-ad10-cd94d2d8a7fb.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1402.370767] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6308358b-8a38-4242-ad10-cd94d2d8a7fb/6308358b-8a38-4242-ad10-cd94d2d8a7fb.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1402.370951] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1402.371227] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-17b8cf1b-f8bc-436e-b31d-230824f74b7f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1402.380689] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1402.380862] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1402.381547] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf72626f-6281-4d27-ac5f-57cc388d9ef9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1402.386219] env[61868]: DEBUG oslo_vmware.api [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1402.386219] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52a1e134-e4ef-60e4-7402-41a416aef00f" [ 1402.386219] env[61868]: _type = "Task" [ 1402.386219] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1402.393722] env[61868]: DEBUG oslo_vmware.api [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52a1e134-e4ef-60e4-7402-41a416aef00f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1402.429580] env[61868]: DEBUG oslo_vmware.api [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316200, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1402.646425] env[61868]: DEBUG nova.network.neutron [req-fed27476-ba7c-4bfa-b2f2-42760a9b6f48 req-ebf9864b-0c85-4e0c-8df1-9167977037f8 service nova] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Updated VIF entry in instance network info cache for port 22435bb8-30a0-4fbb-b86c-a27082b82e4a. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1402.646796] env[61868]: DEBUG nova.network.neutron [req-fed27476-ba7c-4bfa-b2f2-42760a9b6f48 req-ebf9864b-0c85-4e0c-8df1-9167977037f8 service nova] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Updating instance_info_cache with network_info: [{"id": "22435bb8-30a0-4fbb-b86c-a27082b82e4a", "address": "fa:16:3e:89:f5:a6", "network": {"id": "7e8cbecf-dfdd-43f2-8745-ae1d2d55d309", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-33935587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a7014141b24101ba7920bc83be4a9e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4db2961d-273d-4634-9d06-a94fa9d384fb", "external-id": "nsx-vlan-transportzone-572", "segmentation_id": 572, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22435bb8-30", "ovs_interfaceid": "22435bb8-30a0-4fbb-b86c-a27082b82e4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1402.805309] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 7dc60bc8-86bc-4c2f-80c2-56c59b6fb96e] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1402.896330] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Preparing fetch location {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1402.896541] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Fetch image to [datastore1] OSTACK_IMG_282f16e7-a975-439f-88ba-2b80c05f32f4/OSTACK_IMG_282f16e7-a975-439f-88ba-2b80c05f32f4.vmdk {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1402.896726] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Downloading stream optimized image 6308358b-8a38-4242-ad10-cd94d2d8a7fb to [datastore1] OSTACK_IMG_282f16e7-a975-439f-88ba-2b80c05f32f4/OSTACK_IMG_282f16e7-a975-439f-88ba-2b80c05f32f4.vmdk on the data store datastore1 as vApp {{(pid=61868) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1402.896932] env[61868]: DEBUG nova.virt.vmwareapi.images [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Downloading image file data 6308358b-8a38-4242-ad10-cd94d2d8a7fb to the ESX as VM named 'OSTACK_IMG_282f16e7-a975-439f-88ba-2b80c05f32f4' {{(pid=61868) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1402.941035] env[61868]: DEBUG oslo_vmware.api [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316200, 'name': PowerOnVM_Task, 'duration_secs': 0.646828} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1402.941225] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1402.944049] env[61868]: DEBUG nova.compute.manager [None req-2f075402-4dd8-4a55-a049-87299fda2571 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1402.944835] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-478abb73-3855-4978-8200-eceda1244548 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1402.967869] env[61868]: DEBUG oslo_vmware.rw_handles [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1402.967869] env[61868]: value = "resgroup-9" [ 1402.967869] env[61868]: _type = "ResourcePool" [ 1402.967869] env[61868]: }. {{(pid=61868) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1402.968162] env[61868]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-77f92367-2aac-4bb7-a683-21023abdfd26 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1402.988783] env[61868]: DEBUG oslo_vmware.rw_handles [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lease: (returnval){ [ 1402.988783] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52325457-6786-5e17-967e-86812a3790cf" [ 1402.988783] env[61868]: _type = "HttpNfcLease" [ 1402.988783] env[61868]: } obtained for vApp import into resource pool (val){ [ 1402.988783] env[61868]: value = "resgroup-9" [ 1402.988783] env[61868]: _type = "ResourcePool" [ 1402.988783] env[61868]: }. {{(pid=61868) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1402.989169] env[61868]: DEBUG oslo_vmware.api [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the lease: (returnval){ [ 1402.989169] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52325457-6786-5e17-967e-86812a3790cf" [ 1402.989169] env[61868]: _type = "HttpNfcLease" [ 1402.989169] env[61868]: } to be ready. {{(pid=61868) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1402.995044] env[61868]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1402.995044] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52325457-6786-5e17-967e-86812a3790cf" [ 1402.995044] env[61868]: _type = "HttpNfcLease" [ 1402.995044] env[61868]: } is initializing. {{(pid=61868) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1403.149848] env[61868]: DEBUG oslo_concurrency.lockutils [req-fed27476-ba7c-4bfa-b2f2-42760a9b6f48 req-ebf9864b-0c85-4e0c-8df1-9167977037f8 service nova] Releasing lock "refresh_cache-b411db59-3058-4aa8-ac33-d3897c36c123" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1403.308300] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: afd7ff53-f164-4f3d-9ece-28a10245e38c] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1403.498015] env[61868]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1403.498015] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52325457-6786-5e17-967e-86812a3790cf" [ 1403.498015] env[61868]: _type = "HttpNfcLease" [ 1403.498015] env[61868]: } is ready. {{(pid=61868) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1403.499822] env[61868]: DEBUG oslo_vmware.rw_handles [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1403.499822] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52325457-6786-5e17-967e-86812a3790cf" [ 1403.499822] env[61868]: _type = "HttpNfcLease" [ 1403.499822] env[61868]: }. {{(pid=61868) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1403.499822] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f2b9cea-f8dc-4d3e-9109-9d65fe512f22 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1403.507775] env[61868]: DEBUG oslo_vmware.rw_handles [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5203aaa2-893b-1089-ba5e-0d92e9620023/disk-0.vmdk from lease info. {{(pid=61868) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1403.508805] env[61868]: DEBUG oslo_vmware.rw_handles [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Creating HTTP connection to write to file with size = 31668224 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5203aaa2-893b-1089-ba5e-0d92e9620023/disk-0.vmdk. {{(pid=61868) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1403.572105] env[61868]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-4a6439ad-b92c-4da7-8b07-8ebf605d2129 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1403.812106] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 04b0079b-8f5c-4e51-8bb9-52d81273713e] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1404.316060] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: f2d864f1-cd11-4b62-857d-789cf045f22c] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1404.616668] env[61868]: DEBUG oslo_vmware.rw_handles [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Completed reading data from the image iterator. {{(pid=61868) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1404.616997] env[61868]: DEBUG oslo_vmware.rw_handles [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5203aaa2-893b-1089-ba5e-0d92e9620023/disk-0.vmdk. {{(pid=61868) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1404.618131] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cba14b59-5046-4356-98d7-e0faef350e43 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1404.627326] env[61868]: DEBUG oslo_vmware.rw_handles [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5203aaa2-893b-1089-ba5e-0d92e9620023/disk-0.vmdk is in state: ready. {{(pid=61868) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1404.627555] env[61868]: DEBUG oslo_vmware.rw_handles [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5203aaa2-893b-1089-ba5e-0d92e9620023/disk-0.vmdk. {{(pid=61868) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1404.627858] env[61868]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-0bd078c1-23b2-49a1-ac3a-93dff77b139d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1404.816100] env[61868]: DEBUG oslo_vmware.rw_handles [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5203aaa2-893b-1089-ba5e-0d92e9620023/disk-0.vmdk. {{(pid=61868) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1404.816508] env[61868]: INFO nova.virt.vmwareapi.images [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Downloaded image file data 6308358b-8a38-4242-ad10-cd94d2d8a7fb [ 1404.817561] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14c079b8-ce00-4822-aa9f-f8f7d4fedba4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1404.822149] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 59171bf1-f2d4-4ff5-a661-025f5dca7600] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1404.836969] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-33fc83ba-a443-4b2a-a782-e99f3ff48eb6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1404.859660] env[61868]: INFO nova.virt.vmwareapi.images [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] The imported VM was unregistered [ 1404.862148] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Caching image {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1404.862409] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Creating directory with path [datastore1] devstack-image-cache_base/6308358b-8a38-4242-ad10-cd94d2d8a7fb {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1404.862997] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-432e7cb6-5da2-4d15-8999-f9ddd61fa3b2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1404.874313] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Created directory with path [datastore1] devstack-image-cache_base/6308358b-8a38-4242-ad10-cd94d2d8a7fb {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1404.874552] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_282f16e7-a975-439f-88ba-2b80c05f32f4/OSTACK_IMG_282f16e7-a975-439f-88ba-2b80c05f32f4.vmdk to [datastore1] devstack-image-cache_base/6308358b-8a38-4242-ad10-cd94d2d8a7fb/6308358b-8a38-4242-ad10-cd94d2d8a7fb.vmdk. {{(pid=61868) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1404.874843] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-9133d4f7-178e-4fef-9972-4067d4fa69fe {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1404.882225] env[61868]: DEBUG oslo_vmware.api [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1404.882225] env[61868]: value = "task-1316203" [ 1404.882225] env[61868]: _type = "Task" [ 1404.882225] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1404.890235] env[61868]: DEBUG oslo_vmware.api [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316203, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1405.250304] env[61868]: INFO nova.compute.manager [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Rescuing [ 1405.250587] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquiring lock "refresh_cache-28a14ebd-6003-4e20-b556-2a3b939f15c8" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1405.250746] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquired lock "refresh_cache-28a14ebd-6003-4e20-b556-2a3b939f15c8" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1405.250919] env[61868]: DEBUG nova.network.neutron [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1405.325688] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 273a38e7-8eb0-46fc-b9e2-31a9a7fc34db] Instance has had 0 of 5 cleanup attempts {{(pid=61868) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1405.391448] env[61868]: DEBUG oslo_vmware.api [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316203, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1405.829514] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1405.829833] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Cleaning up deleted instances with incomplete migration {{(pid=61868) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1405.892017] env[61868]: DEBUG oslo_vmware.api [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316203, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1405.937295] env[61868]: DEBUG nova.network.neutron [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Updating instance_info_cache with network_info: [{"id": "b1bf2dec-5f14-461f-aa54-f6d67bc0dcb5", "address": "fa:16:3e:5c:1c:e5", "network": {"id": "5be7f7b7-0d32-48b1-979d-5a4a872e5b1a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1791355381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b3904d578a0e47d58eae75b1d4fbf41c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1bf2dec-5f", "ovs_interfaceid": "b1bf2dec-5f14-461f-aa54-f6d67bc0dcb5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1406.393048] env[61868]: DEBUG oslo_vmware.api [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316203, 'name': MoveVirtualDisk_Task} progress is 18%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1406.440028] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Releasing lock "refresh_cache-28a14ebd-6003-4e20-b556-2a3b939f15c8" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1406.895380] env[61868]: DEBUG oslo_vmware.api [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316203, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1406.974099] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1406.974435] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a6d1d0dc-9de4-403c-bcac-c82642f928d7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1406.984017] env[61868]: DEBUG oslo_vmware.api [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1406.984017] env[61868]: value = "task-1316204" [ 1406.984017] env[61868]: _type = "Task" [ 1406.984017] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1406.997401] env[61868]: DEBUG oslo_vmware.api [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316204, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1407.075404] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1407.075704] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1407.075850] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Starting heal instance info cache {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1407.076055] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Rebuilding the list of instances to heal {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1407.395336] env[61868]: DEBUG oslo_vmware.api [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316203, 'name': MoveVirtualDisk_Task} progress is 60%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1407.494282] env[61868]: DEBUG oslo_vmware.api [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316204, 'name': PowerOffVM_Task, 'duration_secs': 0.340925} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1407.494594] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1407.495544] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f002e93-ace9-4da7-9a20-aff201f57696 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1407.516162] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1626f7b-2a45-4b06-9b3d-362b6c763faf {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1407.547814] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1407.548167] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7ce823c7-06d0-45d1-bbd4-8aecdbc497e0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1407.555122] env[61868]: DEBUG oslo_vmware.api [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1407.555122] env[61868]: value = "task-1316205" [ 1407.555122] env[61868]: _type = "Task" [ 1407.555122] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1407.564615] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] VM already powered off {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1407.564840] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1407.565158] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1407.565269] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1407.565459] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1407.565791] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c2ade8e0-5c14-4ad8-9aa7-965338dbb9e5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1407.575839] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1407.576011] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1407.576794] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-86a6a770-310f-450f-9087-e9112c8ff8b9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1407.580578] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "refresh_cache-b411db59-3058-4aa8-ac33-d3897c36c123" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1407.580712] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquired lock "refresh_cache-b411db59-3058-4aa8-ac33-d3897c36c123" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1407.580811] env[61868]: DEBUG nova.network.neutron [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Forcefully refreshing network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1407.580962] env[61868]: DEBUG nova.objects.instance [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lazy-loading 'info_cache' on Instance uuid b411db59-3058-4aa8-ac33-d3897c36c123 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1407.585571] env[61868]: DEBUG oslo_vmware.api [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1407.585571] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]523011e6-6dc8-d337-9d0a-a0ce48b7835a" [ 1407.585571] env[61868]: _type = "Task" [ 1407.585571] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1407.593988] env[61868]: DEBUG oslo_vmware.api [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]523011e6-6dc8-d337-9d0a-a0ce48b7835a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1407.896104] env[61868]: DEBUG oslo_vmware.api [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316203, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1408.096744] env[61868]: DEBUG oslo_vmware.api [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]523011e6-6dc8-d337-9d0a-a0ce48b7835a, 'name': SearchDatastore_Task, 'duration_secs': 0.024933} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1408.098268] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6cdc079d-9a9a-45a9-80d1-dfc571ffba7a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1408.104438] env[61868]: DEBUG oslo_vmware.api [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1408.104438] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]527fbb89-fbf2-4774-1360-d625d781c405" [ 1408.104438] env[61868]: _type = "Task" [ 1408.104438] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1408.112810] env[61868]: DEBUG oslo_vmware.api [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]527fbb89-fbf2-4774-1360-d625d781c405, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1408.397320] env[61868]: DEBUG oslo_vmware.api [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316203, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1408.615047] env[61868]: DEBUG oslo_vmware.api [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]527fbb89-fbf2-4774-1360-d625d781c405, 'name': SearchDatastore_Task, 'duration_secs': 0.025888} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1408.615047] env[61868]: DEBUG oslo_concurrency.lockutils [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1408.615274] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 28a14ebd-6003-4e20-b556-2a3b939f15c8/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d-rescue.vmdk. {{(pid=61868) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1408.615504] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e52a4ac1-c127-4591-a3b9-17b17b92a841 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1408.622326] env[61868]: DEBUG oslo_vmware.api [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1408.622326] env[61868]: value = "task-1316206" [ 1408.622326] env[61868]: _type = "Task" [ 1408.622326] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1408.630704] env[61868]: DEBUG oslo_vmware.api [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316206, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1408.896903] env[61868]: DEBUG oslo_vmware.api [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316203, 'name': MoveVirtualDisk_Task, 'duration_secs': 4.005947} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1408.897281] env[61868]: INFO nova.virt.vmwareapi.ds_util [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_282f16e7-a975-439f-88ba-2b80c05f32f4/OSTACK_IMG_282f16e7-a975-439f-88ba-2b80c05f32f4.vmdk to [datastore1] devstack-image-cache_base/6308358b-8a38-4242-ad10-cd94d2d8a7fb/6308358b-8a38-4242-ad10-cd94d2d8a7fb.vmdk. [ 1408.897417] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Cleaning up location [datastore1] OSTACK_IMG_282f16e7-a975-439f-88ba-2b80c05f32f4 {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1408.897545] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_282f16e7-a975-439f-88ba-2b80c05f32f4 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1408.897845] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ead6f2a7-1b6d-49e7-814c-9b35c3ccf155 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1408.904266] env[61868]: DEBUG oslo_vmware.api [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1408.904266] env[61868]: value = "task-1316207" [ 1408.904266] env[61868]: _type = "Task" [ 1408.904266] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1408.911222] env[61868]: DEBUG oslo_vmware.api [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316207, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1409.132646] env[61868]: DEBUG oslo_vmware.api [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316206, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1409.307714] env[61868]: DEBUG nova.network.neutron [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Updating instance_info_cache with network_info: [{"id": "22435bb8-30a0-4fbb-b86c-a27082b82e4a", "address": "fa:16:3e:89:f5:a6", "network": {"id": "7e8cbecf-dfdd-43f2-8745-ae1d2d55d309", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-33935587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "28a7014141b24101ba7920bc83be4a9e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4db2961d-273d-4634-9d06-a94fa9d384fb", "external-id": "nsx-vlan-transportzone-572", "segmentation_id": 572, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22435bb8-30", "ovs_interfaceid": "22435bb8-30a0-4fbb-b86c-a27082b82e4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1409.413950] env[61868]: DEBUG oslo_vmware.api [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316207, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.346935} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1409.414168] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1409.414337] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6308358b-8a38-4242-ad10-cd94d2d8a7fb/6308358b-8a38-4242-ad10-cd94d2d8a7fb.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1409.414575] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6308358b-8a38-4242-ad10-cd94d2d8a7fb/6308358b-8a38-4242-ad10-cd94d2d8a7fb.vmdk to [datastore1] b411db59-3058-4aa8-ac33-d3897c36c123/b411db59-3058-4aa8-ac33-d3897c36c123.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1409.414824] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-16ee6422-06c5-4ae3-a4ed-675644c70a13 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1409.421695] env[61868]: DEBUG oslo_vmware.api [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1409.421695] env[61868]: value = "task-1316208" [ 1409.421695] env[61868]: _type = "Task" [ 1409.421695] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1409.428832] env[61868]: DEBUG oslo_vmware.api [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316208, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1409.635597] env[61868]: DEBUG oslo_vmware.api [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316206, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1409.810986] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Releasing lock "refresh_cache-b411db59-3058-4aa8-ac33-d3897c36c123" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1409.811329] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Updated the network info_cache for instance {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1409.811573] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1409.811832] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1409.812116] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1409.812436] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1409.813389] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1409.933909] env[61868]: DEBUG oslo_vmware.api [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316208, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1410.134370] env[61868]: DEBUG oslo_vmware.api [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316206, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.377003} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1410.134669] env[61868]: INFO nova.virt.vmwareapi.ds_util [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore2] 28a14ebd-6003-4e20-b556-2a3b939f15c8/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d-rescue.vmdk. [ 1410.135561] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-947b0a77-5534-444e-a0a3-e91a86e2292d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1410.163311] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Reconfiguring VM instance instance-00000070 to attach disk [datastore2] 28a14ebd-6003-4e20-b556-2a3b939f15c8/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d-rescue.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1410.163708] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-093462f8-cf28-434d-84e1-0e58e2d336dc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1410.184453] env[61868]: DEBUG oslo_vmware.api [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1410.184453] env[61868]: value = "task-1316209" [ 1410.184453] env[61868]: _type = "Task" [ 1410.184453] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1410.193295] env[61868]: DEBUG oslo_vmware.api [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316209, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1410.431658] env[61868]: DEBUG oslo_vmware.api [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316208, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1410.696211] env[61868]: DEBUG oslo_vmware.api [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316209, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1410.934618] env[61868]: DEBUG oslo_vmware.api [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316208, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1411.196161] env[61868]: DEBUG oslo_vmware.api [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316209, 'name': ReconfigVM_Task, 'duration_secs': 0.588976} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1411.196438] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Reconfigured VM instance instance-00000070 to attach disk [datastore2] 28a14ebd-6003-4e20-b556-2a3b939f15c8/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d-rescue.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1411.197392] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5966dca-66e6-42a0-a1bb-54288a50f358 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1411.226438] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a632a0ff-f63e-47cd-b1f7-3fdaede05972 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1411.243974] env[61868]: DEBUG oslo_vmware.api [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1411.243974] env[61868]: value = "task-1316210" [ 1411.243974] env[61868]: _type = "Task" [ 1411.243974] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1411.252773] env[61868]: DEBUG oslo_vmware.api [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316210, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1411.434309] env[61868]: DEBUG oslo_vmware.api [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316208, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1411.753502] env[61868]: DEBUG oslo_vmware.api [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316210, 'name': ReconfigVM_Task, 'duration_secs': 0.490573} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1411.753722] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1411.753981] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-38fb7b1d-b55d-4704-a717-1900e745c5ee {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1411.759297] env[61868]: DEBUG oslo_vmware.api [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1411.759297] env[61868]: value = "task-1316211" [ 1411.759297] env[61868]: _type = "Task" [ 1411.759297] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1411.766586] env[61868]: DEBUG oslo_vmware.api [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316211, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1411.933299] env[61868]: DEBUG oslo_vmware.api [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316208, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.486333} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1411.933606] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6308358b-8a38-4242-ad10-cd94d2d8a7fb/6308358b-8a38-4242-ad10-cd94d2d8a7fb.vmdk to [datastore1] b411db59-3058-4aa8-ac33-d3897c36c123/b411db59-3058-4aa8-ac33-d3897c36c123.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1411.934478] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dca3a937-765e-44a1-91cd-cf3df5088900 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1411.956237] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] b411db59-3058-4aa8-ac33-d3897c36c123/b411db59-3058-4aa8-ac33-d3897c36c123.vmdk or device None with type streamOptimized {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1411.956613] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2a7d1c08-146f-496b-9fd6-1e7b9fa092a2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1411.975089] env[61868]: DEBUG oslo_vmware.api [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1411.975089] env[61868]: value = "task-1316212" [ 1411.975089] env[61868]: _type = "Task" [ 1411.975089] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1411.982629] env[61868]: DEBUG oslo_vmware.api [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316212, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1412.270373] env[61868]: DEBUG oslo_vmware.api [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316211, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1412.484088] env[61868]: DEBUG oslo_vmware.api [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316212, 'name': ReconfigVM_Task, 'duration_secs': 0.318471} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1412.484375] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Reconfigured VM instance instance-0000006f to attach disk [datastore1] b411db59-3058-4aa8-ac33-d3897c36c123/b411db59-3058-4aa8-ac33-d3897c36c123.vmdk or device None with type streamOptimized {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1412.484995] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a46d3165-edd6-4c30-8bac-a1d08ab50d2d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1412.491161] env[61868]: DEBUG oslo_vmware.api [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1412.491161] env[61868]: value = "task-1316213" [ 1412.491161] env[61868]: _type = "Task" [ 1412.491161] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1412.498389] env[61868]: DEBUG oslo_vmware.api [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316213, 'name': Rename_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1412.769734] env[61868]: DEBUG oslo_vmware.api [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316211, 'name': PowerOnVM_Task, 'duration_secs': 0.588626} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1412.769975] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1412.772684] env[61868]: DEBUG nova.compute.manager [None req-e2cb3510-d3ac-4048-bfa7-06b4d79cb3ee tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1412.773414] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3801dc86-dae0-4c0b-9822-8e09c5262d79 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.001320] env[61868]: DEBUG oslo_vmware.api [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316213, 'name': Rename_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1413.501468] env[61868]: DEBUG oslo_vmware.api [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316213, 'name': Rename_Task} progress is 99%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1414.002281] env[61868]: DEBUG oslo_vmware.api [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316213, 'name': Rename_Task, 'duration_secs': 1.135798} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1414.002569] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1414.002812] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-54de6895-ae9f-4705-9553-af381912a965 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1414.009188] env[61868]: DEBUG oslo_vmware.api [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1414.009188] env[61868]: value = "task-1316214" [ 1414.009188] env[61868]: _type = "Task" [ 1414.009188] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1414.016310] env[61868]: DEBUG oslo_vmware.api [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316214, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1414.325873] env[61868]: INFO nova.compute.manager [None req-1e144cd1-3474-4f87-9e32-227ebb8ba9af tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Unrescuing [ 1414.326185] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1e144cd1-3474-4f87-9e32-227ebb8ba9af tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquiring lock "refresh_cache-28a14ebd-6003-4e20-b556-2a3b939f15c8" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1414.326342] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1e144cd1-3474-4f87-9e32-227ebb8ba9af tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquired lock "refresh_cache-28a14ebd-6003-4e20-b556-2a3b939f15c8" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1414.326511] env[61868]: DEBUG nova.network.neutron [None req-1e144cd1-3474-4f87-9e32-227ebb8ba9af tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1414.518987] env[61868]: DEBUG oslo_vmware.api [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316214, 'name': PowerOnVM_Task, 'duration_secs': 0.419636} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1414.519256] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1414.614268] env[61868]: DEBUG nova.compute.manager [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1414.615193] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-728158de-7959-4738-ae8c-cbfdca409892 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1415.034249] env[61868]: DEBUG nova.network.neutron [None req-1e144cd1-3474-4f87-9e32-227ebb8ba9af tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Updating instance_info_cache with network_info: [{"id": "b1bf2dec-5f14-461f-aa54-f6d67bc0dcb5", "address": "fa:16:3e:5c:1c:e5", "network": {"id": "5be7f7b7-0d32-48b1-979d-5a4a872e5b1a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1791355381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b3904d578a0e47d58eae75b1d4fbf41c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1bf2dec-5f", "ovs_interfaceid": "b1bf2dec-5f14-461f-aa54-f6d67bc0dcb5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1415.133720] env[61868]: DEBUG oslo_concurrency.lockutils [None req-eb6389d3-69be-43d6-87ce-0420cc01b25e tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "b411db59-3058-4aa8-ac33-d3897c36c123" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 20.954s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1415.537612] env[61868]: DEBUG oslo_concurrency.lockutils [None req-1e144cd1-3474-4f87-9e32-227ebb8ba9af tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Releasing lock "refresh_cache-28a14ebd-6003-4e20-b556-2a3b939f15c8" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1415.538248] env[61868]: DEBUG nova.objects.instance [None req-1e144cd1-3474-4f87-9e32-227ebb8ba9af tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lazy-loading 'flavor' on Instance uuid 28a14ebd-6003-4e20-b556-2a3b939f15c8 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1416.043743] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d914c821-b9c0-42a7-ae1c-938c820aaf0b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1416.064849] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a4c10389-bfea-4739-9198-6a9ca4f133f7 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "b411db59-3058-4aa8-ac33-d3897c36c123" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1416.065102] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a4c10389-bfea-4739-9198-6a9ca4f133f7 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "b411db59-3058-4aa8-ac33-d3897c36c123" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1416.065308] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a4c10389-bfea-4739-9198-6a9ca4f133f7 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "b411db59-3058-4aa8-ac33-d3897c36c123-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1416.065493] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a4c10389-bfea-4739-9198-6a9ca4f133f7 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "b411db59-3058-4aa8-ac33-d3897c36c123-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1416.065676] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a4c10389-bfea-4739-9198-6a9ca4f133f7 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "b411db59-3058-4aa8-ac33-d3897c36c123-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1416.067220] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e144cd1-3474-4f87-9e32-227ebb8ba9af tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1416.067490] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bb6c399f-14dc-465b-9ee6-b407d9adf4d1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1416.069168] env[61868]: INFO nova.compute.manager [None req-a4c10389-bfea-4739-9198-6a9ca4f133f7 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Terminating instance [ 1416.071077] env[61868]: DEBUG nova.compute.manager [None req-a4c10389-bfea-4739-9198-6a9ca4f133f7 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1416.071226] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a4c10389-bfea-4739-9198-6a9ca4f133f7 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1416.071917] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49d05f4a-8d28-4b77-b4c9-8e09c58bbb00 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1416.075775] env[61868]: DEBUG oslo_vmware.api [None req-1e144cd1-3474-4f87-9e32-227ebb8ba9af tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1416.075775] env[61868]: value = "task-1316215" [ 1416.075775] env[61868]: _type = "Task" [ 1416.075775] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1416.080721] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4c10389-bfea-4739-9198-6a9ca4f133f7 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1416.081251] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-da554e39-d390-4069-a11d-56564a0b86c7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1416.085961] env[61868]: DEBUG oslo_vmware.api [None req-1e144cd1-3474-4f87-9e32-227ebb8ba9af tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316215, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1416.087462] env[61868]: DEBUG oslo_vmware.api [None req-a4c10389-bfea-4739-9198-6a9ca4f133f7 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1416.087462] env[61868]: value = "task-1316216" [ 1416.087462] env[61868]: _type = "Task" [ 1416.087462] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1416.095220] env[61868]: DEBUG oslo_vmware.api [None req-a4c10389-bfea-4739-9198-6a9ca4f133f7 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316216, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1416.585576] env[61868]: DEBUG oslo_vmware.api [None req-1e144cd1-3474-4f87-9e32-227ebb8ba9af tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316215, 'name': PowerOffVM_Task, 'duration_secs': 0.213669} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1416.585830] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e144cd1-3474-4f87-9e32-227ebb8ba9af tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1416.590990] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-1e144cd1-3474-4f87-9e32-227ebb8ba9af tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Reconfiguring VM instance instance-00000070 to detach disk 2001 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1416.591616] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-95b2c0e0-f74b-43d6-94cd-94b13232af86 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1416.611174] env[61868]: DEBUG oslo_vmware.api [None req-a4c10389-bfea-4739-9198-6a9ca4f133f7 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316216, 'name': PowerOffVM_Task, 'duration_secs': 0.170313} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1416.612196] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4c10389-bfea-4739-9198-6a9ca4f133f7 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1416.612373] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a4c10389-bfea-4739-9198-6a9ca4f133f7 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1416.612652] env[61868]: DEBUG oslo_vmware.api [None req-1e144cd1-3474-4f87-9e32-227ebb8ba9af tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1416.612652] env[61868]: value = "task-1316217" [ 1416.612652] env[61868]: _type = "Task" [ 1416.612652] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1416.612850] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e53a1dbe-7bce-431c-88c6-5a099d9013ce {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1416.621627] env[61868]: DEBUG oslo_vmware.api [None req-1e144cd1-3474-4f87-9e32-227ebb8ba9af tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316217, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1416.678126] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a4c10389-bfea-4739-9198-6a9ca4f133f7 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1416.678378] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a4c10389-bfea-4739-9198-6a9ca4f133f7 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1416.678586] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4c10389-bfea-4739-9198-6a9ca4f133f7 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Deleting the datastore file [datastore1] b411db59-3058-4aa8-ac33-d3897c36c123 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1416.678888] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1773697d-2f12-4dab-bf93-9b625c4ee62b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1416.684944] env[61868]: DEBUG oslo_vmware.api [None req-a4c10389-bfea-4739-9198-6a9ca4f133f7 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for the task: (returnval){ [ 1416.684944] env[61868]: value = "task-1316219" [ 1416.684944] env[61868]: _type = "Task" [ 1416.684944] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1416.693100] env[61868]: DEBUG oslo_vmware.api [None req-a4c10389-bfea-4739-9198-6a9ca4f133f7 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316219, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1417.123324] env[61868]: DEBUG oslo_vmware.api [None req-1e144cd1-3474-4f87-9e32-227ebb8ba9af tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316217, 'name': ReconfigVM_Task, 'duration_secs': 0.222742} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1417.123685] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-1e144cd1-3474-4f87-9e32-227ebb8ba9af tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Reconfigured VM instance instance-00000070 to detach disk 2001 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1417.123784] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e144cd1-3474-4f87-9e32-227ebb8ba9af tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1417.124056] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-43e722ec-78c7-4dbf-a3c8-dd8f8e613ee2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1417.129629] env[61868]: DEBUG oslo_vmware.api [None req-1e144cd1-3474-4f87-9e32-227ebb8ba9af tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1417.129629] env[61868]: value = "task-1316220" [ 1417.129629] env[61868]: _type = "Task" [ 1417.129629] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1417.138394] env[61868]: DEBUG oslo_vmware.api [None req-1e144cd1-3474-4f87-9e32-227ebb8ba9af tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316220, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1417.193971] env[61868]: DEBUG oslo_vmware.api [None req-a4c10389-bfea-4739-9198-6a9ca4f133f7 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Task: {'id': task-1316219, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152509} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1417.194249] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4c10389-bfea-4739-9198-6a9ca4f133f7 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1417.194440] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a4c10389-bfea-4739-9198-6a9ca4f133f7 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1417.194616] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-a4c10389-bfea-4739-9198-6a9ca4f133f7 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1417.194791] env[61868]: INFO nova.compute.manager [None req-a4c10389-bfea-4739-9198-6a9ca4f133f7 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1417.195047] env[61868]: DEBUG oslo.service.loopingcall [None req-a4c10389-bfea-4739-9198-6a9ca4f133f7 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1417.195251] env[61868]: DEBUG nova.compute.manager [-] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1417.195342] env[61868]: DEBUG nova.network.neutron [-] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1417.642753] env[61868]: DEBUG oslo_vmware.api [None req-1e144cd1-3474-4f87-9e32-227ebb8ba9af tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316220, 'name': PowerOnVM_Task, 'duration_secs': 0.478799} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1417.643057] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e144cd1-3474-4f87-9e32-227ebb8ba9af tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1417.643297] env[61868]: DEBUG nova.compute.manager [None req-1e144cd1-3474-4f87-9e32-227ebb8ba9af tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1417.644061] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feab16e7-7cbe-4782-befd-5c14baf7320e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1417.678491] env[61868]: DEBUG nova.compute.manager [req-d152eebb-71b2-4ee2-9929-6df07220adb6 req-7e9b7886-4abc-4636-91a9-a40f19be9dc4 service nova] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Received event network-vif-deleted-22435bb8-30a0-4fbb-b86c-a27082b82e4a {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1417.678576] env[61868]: INFO nova.compute.manager [req-d152eebb-71b2-4ee2-9929-6df07220adb6 req-7e9b7886-4abc-4636-91a9-a40f19be9dc4 service nova] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Neutron deleted interface 22435bb8-30a0-4fbb-b86c-a27082b82e4a; detaching it from the instance and deleting it from the info cache [ 1417.678788] env[61868]: DEBUG nova.network.neutron [req-d152eebb-71b2-4ee2-9929-6df07220adb6 req-7e9b7886-4abc-4636-91a9-a40f19be9dc4 service nova] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1418.160096] env[61868]: DEBUG nova.network.neutron [-] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1418.181223] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bc0d98dd-70c6-4cbd-98cf-84d4a7c3b39a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1418.190672] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cdca65f-2986-4f81-b608-87fab89b20a2 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1418.214030] env[61868]: DEBUG nova.compute.manager [req-d152eebb-71b2-4ee2-9929-6df07220adb6 req-7e9b7886-4abc-4636-91a9-a40f19be9dc4 service nova] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Detach interface failed, port_id=22435bb8-30a0-4fbb-b86c-a27082b82e4a, reason: Instance b411db59-3058-4aa8-ac33-d3897c36c123 could not be found. {{(pid=61868) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1418.662756] env[61868]: INFO nova.compute.manager [-] [instance: b411db59-3058-4aa8-ac33-d3897c36c123] Took 1.47 seconds to deallocate network for instance. [ 1419.169330] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a4c10389-bfea-4739-9198-6a9ca4f133f7 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1419.169689] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a4c10389-bfea-4739-9198-6a9ca4f133f7 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1419.169946] env[61868]: DEBUG nova.objects.instance [None req-a4c10389-bfea-4739-9198-6a9ca4f133f7 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lazy-loading 'resources' on Instance uuid b411db59-3058-4aa8-ac33-d3897c36c123 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1419.731184] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d59ee74a-695d-45d8-aaa4-f8a14fecfbb8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1419.739419] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9621bdc-eabc-4114-8442-8acf94c5b433 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1419.769557] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cac712e6-cccf-4e2f-a382-817c5e556f9f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1419.776770] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f74f418a-b253-491a-9b3e-14db34b352d3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1419.789701] env[61868]: DEBUG nova.compute.provider_tree [None req-a4c10389-bfea-4739-9198-6a9ca4f133f7 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1420.293323] env[61868]: DEBUG nova.scheduler.client.report [None req-a4c10389-bfea-4739-9198-6a9ca4f133f7 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1420.798337] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a4c10389-bfea-4739-9198-6a9ca4f133f7 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.628s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1420.814541] env[61868]: INFO nova.scheduler.client.report [None req-a4c10389-bfea-4739-9198-6a9ca4f133f7 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Deleted allocations for instance b411db59-3058-4aa8-ac33-d3897c36c123 [ 1421.033765] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquiring lock "812b40f7-3b7e-43a5-92ae-df3891e74e4c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1421.034024] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lock "812b40f7-3b7e-43a5-92ae-df3891e74e4c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1421.322726] env[61868]: DEBUG oslo_concurrency.lockutils [None req-a4c10389-bfea-4739-9198-6a9ca4f133f7 tempest-ServerActionsTestOtherB-617170686 tempest-ServerActionsTestOtherB-617170686-project-member] Lock "b411db59-3058-4aa8-ac33-d3897c36c123" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.257s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1421.536883] env[61868]: DEBUG nova.compute.manager [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Starting instance... {{(pid=61868) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1422.120250] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1422.120686] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1422.123138] env[61868]: INFO nova.compute.claims [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1423.146265] env[61868]: DEBUG nova.scheduler.client.report [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Refreshing inventories for resource provider 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1423.159976] env[61868]: DEBUG nova.scheduler.client.report [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Updating ProviderTree inventory for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1423.160218] env[61868]: DEBUG nova.compute.provider_tree [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Updating inventory in ProviderTree for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1423.170030] env[61868]: DEBUG nova.scheduler.client.report [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Refreshing aggregate associations for resource provider 83e67721-2ac9-4a23-aa31-82aca86979c8, aggregates: None {{(pid=61868) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1423.186405] env[61868]: DEBUG nova.scheduler.client.report [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Refreshing trait associations for resource provider 83e67721-2ac9-4a23-aa31-82aca86979c8, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61868) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1423.230347] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94be0578-2548-4c3d-87e3-c7959e4e79cc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1423.237836] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6323e5b-76e9-4663-a318-dbb8793d3ad1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1423.266531] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27180460-4861-40b8-b9c3-1fe64fc7b794 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1423.273437] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db1441c1-45f2-422a-af4f-735cd9abcb1a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1423.286119] env[61868]: DEBUG nova.compute.provider_tree [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1423.788980] env[61868]: DEBUG nova.scheduler.client.report [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1424.294186] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.173s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1424.294714] env[61868]: DEBUG nova.compute.manager [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Start building networks asynchronously for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1424.799901] env[61868]: DEBUG nova.compute.utils [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1424.801437] env[61868]: DEBUG nova.compute.manager [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Allocating IP information in the background. {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1424.801642] env[61868]: DEBUG nova.network.neutron [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] allocate_for_instance() {{(pid=61868) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1424.845247] env[61868]: DEBUG nova.policy [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4dbe667c15f14d13b6ef69bfe770ac70', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b3904d578a0e47d58eae75b1d4fbf41c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61868) authorize /opt/stack/nova/nova/policy.py:201}} [ 1425.108214] env[61868]: DEBUG nova.network.neutron [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Successfully created port: 5aa4db9f-d6ec-47fa-b327-8ab288ed9255 {{(pid=61868) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1425.304928] env[61868]: DEBUG nova.compute.manager [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Start building block device mappings for instance. {{(pid=61868) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1426.319235] env[61868]: DEBUG nova.compute.manager [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Start spawning the instance on the hypervisor. {{(pid=61868) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1426.347740] env[61868]: DEBUG nova.virt.hardware [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-18T17:17:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-18T17:17:10Z,direct_url=,disk_format='vmdk',id=c4fdfa42-c9a0-41ca-801e-4e55c3d4328d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cfcd340b6ca141a38b1f777501f7d3e7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-18T17:17:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1426.348183] env[61868]: DEBUG nova.virt.hardware [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Flavor limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1426.348432] env[61868]: DEBUG nova.virt.hardware [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Image limits 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1426.348803] env[61868]: DEBUG nova.virt.hardware [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Flavor pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1426.348941] env[61868]: DEBUG nova.virt.hardware [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Image pref 0:0:0 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1426.349194] env[61868]: DEBUG nova.virt.hardware [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61868) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1426.349511] env[61868]: DEBUG nova.virt.hardware [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1426.349760] env[61868]: DEBUG nova.virt.hardware [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1426.350028] env[61868]: DEBUG nova.virt.hardware [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Got 1 possible topologies {{(pid=61868) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1426.350380] env[61868]: DEBUG nova.virt.hardware [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1426.350661] env[61868]: DEBUG nova.virt.hardware [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61868) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1426.351952] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86e588f0-1912-47f9-bfb9-b0d34a5521c5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1426.362678] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fb7b9df-fe8d-4d76-972b-c93d961c2338 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1426.490209] env[61868]: DEBUG nova.compute.manager [req-1bbc39a9-fe6a-4545-855a-b2f503dfa802 req-0634d915-df6d-41e1-b8d9-1e02b93337ae service nova] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Received event network-vif-plugged-5aa4db9f-d6ec-47fa-b327-8ab288ed9255 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1426.490209] env[61868]: DEBUG oslo_concurrency.lockutils [req-1bbc39a9-fe6a-4545-855a-b2f503dfa802 req-0634d915-df6d-41e1-b8d9-1e02b93337ae service nova] Acquiring lock "812b40f7-3b7e-43a5-92ae-df3891e74e4c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1426.490209] env[61868]: DEBUG oslo_concurrency.lockutils [req-1bbc39a9-fe6a-4545-855a-b2f503dfa802 req-0634d915-df6d-41e1-b8d9-1e02b93337ae service nova] Lock "812b40f7-3b7e-43a5-92ae-df3891e74e4c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1426.490437] env[61868]: DEBUG oslo_concurrency.lockutils [req-1bbc39a9-fe6a-4545-855a-b2f503dfa802 req-0634d915-df6d-41e1-b8d9-1e02b93337ae service nova] Lock "812b40f7-3b7e-43a5-92ae-df3891e74e4c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1426.490779] env[61868]: DEBUG nova.compute.manager [req-1bbc39a9-fe6a-4545-855a-b2f503dfa802 req-0634d915-df6d-41e1-b8d9-1e02b93337ae service nova] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] No waiting events found dispatching network-vif-plugged-5aa4db9f-d6ec-47fa-b327-8ab288ed9255 {{(pid=61868) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1426.490995] env[61868]: WARNING nova.compute.manager [req-1bbc39a9-fe6a-4545-855a-b2f503dfa802 req-0634d915-df6d-41e1-b8d9-1e02b93337ae service nova] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Received unexpected event network-vif-plugged-5aa4db9f-d6ec-47fa-b327-8ab288ed9255 for instance with vm_state building and task_state spawning. [ 1426.581358] env[61868]: DEBUG nova.network.neutron [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Successfully updated port: 5aa4db9f-d6ec-47fa-b327-8ab288ed9255 {{(pid=61868) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1427.084260] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquiring lock "refresh_cache-812b40f7-3b7e-43a5-92ae-df3891e74e4c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1427.084413] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquired lock "refresh_cache-812b40f7-3b7e-43a5-92ae-df3891e74e4c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1427.084559] env[61868]: DEBUG nova.network.neutron [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1427.614133] env[61868]: DEBUG nova.network.neutron [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Instance cache missing network info. {{(pid=61868) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1427.742325] env[61868]: DEBUG nova.network.neutron [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Updating instance_info_cache with network_info: [{"id": "5aa4db9f-d6ec-47fa-b327-8ab288ed9255", "address": "fa:16:3e:23:f9:38", "network": {"id": "5be7f7b7-0d32-48b1-979d-5a4a872e5b1a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1791355381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b3904d578a0e47d58eae75b1d4fbf41c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5aa4db9f-d6", "ovs_interfaceid": "5aa4db9f-d6ec-47fa-b327-8ab288ed9255", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1428.245462] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Releasing lock "refresh_cache-812b40f7-3b7e-43a5-92ae-df3891e74e4c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1428.245742] env[61868]: DEBUG nova.compute.manager [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Instance network_info: |[{"id": "5aa4db9f-d6ec-47fa-b327-8ab288ed9255", "address": "fa:16:3e:23:f9:38", "network": {"id": "5be7f7b7-0d32-48b1-979d-5a4a872e5b1a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1791355381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b3904d578a0e47d58eae75b1d4fbf41c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5aa4db9f-d6", "ovs_interfaceid": "5aa4db9f-d6ec-47fa-b327-8ab288ed9255", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61868) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1428.246220] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:23:f9:38', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a8b99a46-3e7f-4ef1-9e45-58e6cd17f210', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5aa4db9f-d6ec-47fa-b327-8ab288ed9255', 'vif_model': 'vmxnet3'}] {{(pid=61868) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1428.253751] env[61868]: DEBUG oslo.service.loopingcall [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1428.253965] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Creating VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1428.254207] env[61868]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f672f141-a5d3-4c4b-af9b-0233356414a7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1428.274659] env[61868]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1428.274659] env[61868]: value = "task-1316222" [ 1428.274659] env[61868]: _type = "Task" [ 1428.274659] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1428.285286] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1316222, 'name': CreateVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1428.514747] env[61868]: DEBUG nova.compute.manager [req-f54a6836-33f1-45d3-bace-dc4fb8db28f3 req-5d0b181e-a506-4c8f-b907-23abc883dd2d service nova] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Received event network-changed-5aa4db9f-d6ec-47fa-b327-8ab288ed9255 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1428.514976] env[61868]: DEBUG nova.compute.manager [req-f54a6836-33f1-45d3-bace-dc4fb8db28f3 req-5d0b181e-a506-4c8f-b907-23abc883dd2d service nova] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Refreshing instance network info cache due to event network-changed-5aa4db9f-d6ec-47fa-b327-8ab288ed9255. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1428.515220] env[61868]: DEBUG oslo_concurrency.lockutils [req-f54a6836-33f1-45d3-bace-dc4fb8db28f3 req-5d0b181e-a506-4c8f-b907-23abc883dd2d service nova] Acquiring lock "refresh_cache-812b40f7-3b7e-43a5-92ae-df3891e74e4c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1428.515361] env[61868]: DEBUG oslo_concurrency.lockutils [req-f54a6836-33f1-45d3-bace-dc4fb8db28f3 req-5d0b181e-a506-4c8f-b907-23abc883dd2d service nova] Acquired lock "refresh_cache-812b40f7-3b7e-43a5-92ae-df3891e74e4c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1428.515522] env[61868]: DEBUG nova.network.neutron [req-f54a6836-33f1-45d3-bace-dc4fb8db28f3 req-5d0b181e-a506-4c8f-b907-23abc883dd2d service nova] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Refreshing network info cache for port 5aa4db9f-d6ec-47fa-b327-8ab288ed9255 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1428.785299] env[61868]: DEBUG oslo_vmware.api [-] Task: {'id': task-1316222, 'name': CreateVM_Task, 'duration_secs': 0.285825} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1428.785653] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Created VM on the ESX host {{(pid=61868) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1428.792317] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1428.792496] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1428.792825] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1428.793147] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30b92431-e86f-42b2-b48c-d0042842ab15 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1428.797614] env[61868]: DEBUG oslo_vmware.api [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1428.797614] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52ca63e3-db37-e013-c2e0-4035b5f17c28" [ 1428.797614] env[61868]: _type = "Task" [ 1428.797614] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1428.806442] env[61868]: DEBUG oslo_vmware.api [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52ca63e3-db37-e013-c2e0-4035b5f17c28, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1429.203743] env[61868]: DEBUG nova.network.neutron [req-f54a6836-33f1-45d3-bace-dc4fb8db28f3 req-5d0b181e-a506-4c8f-b907-23abc883dd2d service nova] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Updated VIF entry in instance network info cache for port 5aa4db9f-d6ec-47fa-b327-8ab288ed9255. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1429.204156] env[61868]: DEBUG nova.network.neutron [req-f54a6836-33f1-45d3-bace-dc4fb8db28f3 req-5d0b181e-a506-4c8f-b907-23abc883dd2d service nova] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Updating instance_info_cache with network_info: [{"id": "5aa4db9f-d6ec-47fa-b327-8ab288ed9255", "address": "fa:16:3e:23:f9:38", "network": {"id": "5be7f7b7-0d32-48b1-979d-5a4a872e5b1a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1791355381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b3904d578a0e47d58eae75b1d4fbf41c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5aa4db9f-d6", "ovs_interfaceid": "5aa4db9f-d6ec-47fa-b327-8ab288ed9255", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1429.308285] env[61868]: DEBUG oslo_vmware.api [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52ca63e3-db37-e013-c2e0-4035b5f17c28, 'name': SearchDatastore_Task, 'duration_secs': 0.009564} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1429.308584] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1429.308823] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1429.309069] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1429.309229] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1429.309411] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1429.309666] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-01381c52-8ee6-4a24-a5b8-46ae3c467a5d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1429.318158] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1429.318340] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1429.319014] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-172b1c37-bb43-477f-bd96-4337467911b0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1429.324247] env[61868]: DEBUG oslo_vmware.api [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1429.324247] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52b46a2d-fb57-3268-4203-8e8736db2cce" [ 1429.324247] env[61868]: _type = "Task" [ 1429.324247] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1429.331239] env[61868]: DEBUG oslo_vmware.api [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52b46a2d-fb57-3268-4203-8e8736db2cce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1429.706795] env[61868]: DEBUG oslo_concurrency.lockutils [req-f54a6836-33f1-45d3-bace-dc4fb8db28f3 req-5d0b181e-a506-4c8f-b907-23abc883dd2d service nova] Releasing lock "refresh_cache-812b40f7-3b7e-43a5-92ae-df3891e74e4c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1429.834576] env[61868]: DEBUG oslo_vmware.api [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52b46a2d-fb57-3268-4203-8e8736db2cce, 'name': SearchDatastore_Task, 'duration_secs': 0.008581} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1429.835304] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f5bf15e4-4193-40a2-91bf-d8149d9b864c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1429.840327] env[61868]: DEBUG oslo_vmware.api [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1429.840327] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52553645-761b-fcdc-6f53-3af2988b29f4" [ 1429.840327] env[61868]: _type = "Task" [ 1429.840327] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1429.847534] env[61868]: DEBUG oslo_vmware.api [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52553645-761b-fcdc-6f53-3af2988b29f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1430.350275] env[61868]: DEBUG oslo_vmware.api [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52553645-761b-fcdc-6f53-3af2988b29f4, 'name': SearchDatastore_Task, 'duration_secs': 0.011299} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1430.350534] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1430.350796] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 812b40f7-3b7e-43a5-92ae-df3891e74e4c/812b40f7-3b7e-43a5-92ae-df3891e74e4c.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1430.351116] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7ef73a83-7564-49ba-b88e-64d3fee42460 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1430.357987] env[61868]: DEBUG oslo_vmware.api [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1430.357987] env[61868]: value = "task-1316223" [ 1430.357987] env[61868]: _type = "Task" [ 1430.357987] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1430.365191] env[61868]: DEBUG oslo_vmware.api [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316223, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1430.868112] env[61868]: DEBUG oslo_vmware.api [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316223, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.461134} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1430.868454] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 812b40f7-3b7e-43a5-92ae-df3891e74e4c/812b40f7-3b7e-43a5-92ae-df3891e74e4c.vmdk {{(pid=61868) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1430.868540] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Extending root virtual disk to 1048576 {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1430.868783] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f278466f-ac46-4b63-9ed7-312e44067738 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1430.874727] env[61868]: DEBUG oslo_vmware.api [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1430.874727] env[61868]: value = "task-1316224" [ 1430.874727] env[61868]: _type = "Task" [ 1430.874727] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1430.882711] env[61868]: DEBUG oslo_vmware.api [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316224, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1431.384577] env[61868]: DEBUG oslo_vmware.api [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316224, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.055945} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1431.384939] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Extended root virtual disk {{(pid=61868) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1431.385736] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ec5f24f-3e04-4b4c-8f00-a9fd47152973 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1431.406796] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Reconfiguring VM instance instance-00000072 to attach disk [datastore1] 812b40f7-3b7e-43a5-92ae-df3891e74e4c/812b40f7-3b7e-43a5-92ae-df3891e74e4c.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1431.407051] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-62a3761e-72ca-4c2f-b987-6ea2114ea1d3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1431.426179] env[61868]: DEBUG oslo_vmware.api [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1431.426179] env[61868]: value = "task-1316225" [ 1431.426179] env[61868]: _type = "Task" [ 1431.426179] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1431.433397] env[61868]: DEBUG oslo_vmware.api [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316225, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1431.936511] env[61868]: DEBUG oslo_vmware.api [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316225, 'name': ReconfigVM_Task, 'duration_secs': 0.252101} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1431.937025] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Reconfigured VM instance instance-00000072 to attach disk [datastore1] 812b40f7-3b7e-43a5-92ae-df3891e74e4c/812b40f7-3b7e-43a5-92ae-df3891e74e4c.vmdk or device None with type sparse {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1431.937436] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8dda20bb-cba9-4bea-bff9-d7f301654f55 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1431.944211] env[61868]: DEBUG oslo_vmware.api [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1431.944211] env[61868]: value = "task-1316226" [ 1431.944211] env[61868]: _type = "Task" [ 1431.944211] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1431.953088] env[61868]: DEBUG oslo_vmware.api [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316226, 'name': Rename_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1432.453592] env[61868]: DEBUG oslo_vmware.api [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316226, 'name': Rename_Task, 'duration_secs': 0.180637} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1432.453877] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1432.454136] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-293f3170-2696-42f2-961a-f9f290665c45 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1432.460998] env[61868]: DEBUG oslo_vmware.api [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1432.460998] env[61868]: value = "task-1316227" [ 1432.460998] env[61868]: _type = "Task" [ 1432.460998] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1432.468074] env[61868]: DEBUG oslo_vmware.api [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316227, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1432.971446] env[61868]: DEBUG oslo_vmware.api [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316227, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1433.472016] env[61868]: DEBUG oslo_vmware.api [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316227, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1433.972595] env[61868]: DEBUG oslo_vmware.api [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316227, 'name': PowerOnVM_Task, 'duration_secs': 1.069802} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1433.973111] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1433.973111] env[61868]: INFO nova.compute.manager [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Took 7.65 seconds to spawn the instance on the hypervisor. [ 1433.973349] env[61868]: DEBUG nova.compute.manager [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1433.974175] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2283f3cd-68b5-4ca2-b341-fb23a1d85f40 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1434.490077] env[61868]: INFO nova.compute.manager [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Took 12.45 seconds to build instance. [ 1434.992615] env[61868]: DEBUG oslo_concurrency.lockutils [None req-8312ea1a-76cb-4381-9cd7-3cf3f2e60eca tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lock "812b40f7-3b7e-43a5-92ae-df3891e74e4c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.958s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1435.108386] env[61868]: DEBUG nova.compute.manager [req-6ba2e118-d63c-4fb1-8971-742783129050 req-b413993f-096f-4468-9d11-3f1e8e8d9505 service nova] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Received event network-changed-5aa4db9f-d6ec-47fa-b327-8ab288ed9255 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1435.108553] env[61868]: DEBUG nova.compute.manager [req-6ba2e118-d63c-4fb1-8971-742783129050 req-b413993f-096f-4468-9d11-3f1e8e8d9505 service nova] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Refreshing instance network info cache due to event network-changed-5aa4db9f-d6ec-47fa-b327-8ab288ed9255. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1435.108772] env[61868]: DEBUG oslo_concurrency.lockutils [req-6ba2e118-d63c-4fb1-8971-742783129050 req-b413993f-096f-4468-9d11-3f1e8e8d9505 service nova] Acquiring lock "refresh_cache-812b40f7-3b7e-43a5-92ae-df3891e74e4c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1435.108945] env[61868]: DEBUG oslo_concurrency.lockutils [req-6ba2e118-d63c-4fb1-8971-742783129050 req-b413993f-096f-4468-9d11-3f1e8e8d9505 service nova] Acquired lock "refresh_cache-812b40f7-3b7e-43a5-92ae-df3891e74e4c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1435.109141] env[61868]: DEBUG nova.network.neutron [req-6ba2e118-d63c-4fb1-8971-742783129050 req-b413993f-096f-4468-9d11-3f1e8e8d9505 service nova] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Refreshing network info cache for port 5aa4db9f-d6ec-47fa-b327-8ab288ed9255 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1435.819527] env[61868]: DEBUG nova.network.neutron [req-6ba2e118-d63c-4fb1-8971-742783129050 req-b413993f-096f-4468-9d11-3f1e8e8d9505 service nova] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Updated VIF entry in instance network info cache for port 5aa4db9f-d6ec-47fa-b327-8ab288ed9255. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1435.819901] env[61868]: DEBUG nova.network.neutron [req-6ba2e118-d63c-4fb1-8971-742783129050 req-b413993f-096f-4468-9d11-3f1e8e8d9505 service nova] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Updating instance_info_cache with network_info: [{"id": "5aa4db9f-d6ec-47fa-b327-8ab288ed9255", "address": "fa:16:3e:23:f9:38", "network": {"id": "5be7f7b7-0d32-48b1-979d-5a4a872e5b1a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1791355381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.201", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b3904d578a0e47d58eae75b1d4fbf41c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5aa4db9f-d6", "ovs_interfaceid": "5aa4db9f-d6ec-47fa-b327-8ab288ed9255", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1436.322853] env[61868]: DEBUG oslo_concurrency.lockutils [req-6ba2e118-d63c-4fb1-8971-742783129050 req-b413993f-096f-4468-9d11-3f1e8e8d9505 service nova] Releasing lock "refresh_cache-812b40f7-3b7e-43a5-92ae-df3891e74e4c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1456.579752] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1457.074740] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1457.074935] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61868) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1459.075422] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1459.075766] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager.update_available_resource {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1459.579375] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1459.579644] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1459.579795] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1459.579955] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61868) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1459.580900] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9303a22-3d2a-4bc8-805c-59cf792f1a0e {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1459.589291] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d122bc7-1aec-4055-b1ec-54282d3bc4fc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1459.603995] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d82f0d8e-6583-4a13-bb8f-eea82cadbfab {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1459.610245] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9109e68-071d-442c-9307-0e920991805d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1459.637392] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181310MB free_disk=134GB free_vcpus=48 pci_devices=None {{(pid=61868) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1459.637537] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1459.637706] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1460.665103] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 28a14ebd-6003-4e20-b556-2a3b939f15c8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1460.665382] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 6ce00281-a201-445a-bb75-aac07fd481aa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1460.665382] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 812b40f7-3b7e-43a5-92ae-df3891e74e4c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1460.665565] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=61868) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1460.665710] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=61868) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1460.711167] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d97d9f5c-de2e-4041-8668-4e9a56158df7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1460.718490] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8d19043-51ca-412d-8130-a4d29af0a7e1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1460.748593] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d4c3fc1-1d4f-4ebe-821e-afbc4031e05a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1460.755259] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a61cccb-e37e-474c-897d-35830877787f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1460.768779] env[61868]: DEBUG nova.compute.provider_tree [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1461.271411] env[61868]: DEBUG nova.scheduler.client.report [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1461.776502] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61868) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1461.776735] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.139s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1463.776153] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1463.776517] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1463.776517] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Starting heal instance info cache {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1464.444760] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "refresh_cache-28a14ebd-6003-4e20-b556-2a3b939f15c8" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1464.444910] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquired lock "refresh_cache-28a14ebd-6003-4e20-b556-2a3b939f15c8" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1464.445082] env[61868]: DEBUG nova.network.neutron [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Forcefully refreshing network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1465.635523] env[61868]: DEBUG nova.network.neutron [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Updating instance_info_cache with network_info: [{"id": "b1bf2dec-5f14-461f-aa54-f6d67bc0dcb5", "address": "fa:16:3e:5c:1c:e5", "network": {"id": "5be7f7b7-0d32-48b1-979d-5a4a872e5b1a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1791355381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b3904d578a0e47d58eae75b1d4fbf41c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1bf2dec-5f", "ovs_interfaceid": "b1bf2dec-5f14-461f-aa54-f6d67bc0dcb5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1466.138287] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Releasing lock "refresh_cache-28a14ebd-6003-4e20-b556-2a3b939f15c8" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1466.138591] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Updated the network info_cache for instance {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1466.138865] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1466.139111] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1466.139348] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1471.435445] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1472.242252] env[61868]: DEBUG oslo_concurrency.lockutils [None req-60c93e4e-db83-4dde-ab91-d7984de84f9a tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquiring lock "812b40f7-3b7e-43a5-92ae-df3891e74e4c" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1472.242498] env[61868]: DEBUG oslo_concurrency.lockutils [None req-60c93e4e-db83-4dde-ab91-d7984de84f9a tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lock "812b40f7-3b7e-43a5-92ae-df3891e74e4c" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1472.746071] env[61868]: DEBUG nova.compute.utils [None req-60c93e4e-db83-4dde-ab91-d7984de84f9a tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Using /dev/sd instead of None {{(pid=61868) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1473.249112] env[61868]: DEBUG oslo_concurrency.lockutils [None req-60c93e4e-db83-4dde-ab91-d7984de84f9a tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lock "812b40f7-3b7e-43a5-92ae-df3891e74e4c" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1474.309846] env[61868]: DEBUG oslo_concurrency.lockutils [None req-60c93e4e-db83-4dde-ab91-d7984de84f9a tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquiring lock "812b40f7-3b7e-43a5-92ae-df3891e74e4c" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1474.309846] env[61868]: DEBUG oslo_concurrency.lockutils [None req-60c93e4e-db83-4dde-ab91-d7984de84f9a tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lock "812b40f7-3b7e-43a5-92ae-df3891e74e4c" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1474.309846] env[61868]: INFO nova.compute.manager [None req-60c93e4e-db83-4dde-ab91-d7984de84f9a tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Attaching volume fdc9ac56-be2c-4fc5-a830-1f781e497d60 to /dev/sdb [ 1474.339063] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea4e326e-af02-4c17-b072-793aaffc2663 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1474.346081] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fc2f00d-2804-43c6-abf2-fdac2a80db3d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1474.358709] env[61868]: DEBUG nova.virt.block_device [None req-60c93e4e-db83-4dde-ab91-d7984de84f9a tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Updating existing volume attachment record: ce3fffe4-e900-43f0-a49b-ad8cd3d26c94 {{(pid=61868) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1478.900901] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-60c93e4e-db83-4dde-ab91-d7984de84f9a tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Volume attach. Driver type: vmdk {{(pid=61868) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1478.901255] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-60c93e4e-db83-4dde-ab91-d7984de84f9a tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281660', 'volume_id': 'fdc9ac56-be2c-4fc5-a830-1f781e497d60', 'name': 'volume-fdc9ac56-be2c-4fc5-a830-1f781e497d60', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '812b40f7-3b7e-43a5-92ae-df3891e74e4c', 'attached_at': '', 'detached_at': '', 'volume_id': 'fdc9ac56-be2c-4fc5-a830-1f781e497d60', 'serial': 'fdc9ac56-be2c-4fc5-a830-1f781e497d60'} {{(pid=61868) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1478.902129] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6084a46d-c62f-417c-9aa9-c346188f9987 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1478.918115] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab17386d-95fe-4b66-8ca4-f2e0ab0b5431 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1478.942801] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-60c93e4e-db83-4dde-ab91-d7984de84f9a tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Reconfiguring VM instance instance-00000072 to attach disk [datastore1] volume-fdc9ac56-be2c-4fc5-a830-1f781e497d60/volume-fdc9ac56-be2c-4fc5-a830-1f781e497d60.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1478.943066] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9e0d2037-b784-4c82-ba0b-8bbbf0d5f92f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1478.959943] env[61868]: DEBUG oslo_vmware.api [None req-60c93e4e-db83-4dde-ab91-d7984de84f9a tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1478.959943] env[61868]: value = "task-1316232" [ 1478.959943] env[61868]: _type = "Task" [ 1478.959943] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1478.967114] env[61868]: DEBUG oslo_vmware.api [None req-60c93e4e-db83-4dde-ab91-d7984de84f9a tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316232, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1479.468910] env[61868]: DEBUG oslo_vmware.api [None req-60c93e4e-db83-4dde-ab91-d7984de84f9a tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316232, 'name': ReconfigVM_Task, 'duration_secs': 0.318813} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1479.469393] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-60c93e4e-db83-4dde-ab91-d7984de84f9a tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Reconfigured VM instance instance-00000072 to attach disk [datastore1] volume-fdc9ac56-be2c-4fc5-a830-1f781e497d60/volume-fdc9ac56-be2c-4fc5-a830-1f781e497d60.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1479.473955] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bdd4506a-6e91-4ab8-8a83-764e96bd917f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1479.487550] env[61868]: DEBUG oslo_vmware.api [None req-60c93e4e-db83-4dde-ab91-d7984de84f9a tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1479.487550] env[61868]: value = "task-1316233" [ 1479.487550] env[61868]: _type = "Task" [ 1479.487550] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1479.495992] env[61868]: DEBUG oslo_vmware.api [None req-60c93e4e-db83-4dde-ab91-d7984de84f9a tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316233, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1479.997398] env[61868]: DEBUG oslo_vmware.api [None req-60c93e4e-db83-4dde-ab91-d7984de84f9a tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316233, 'name': ReconfigVM_Task, 'duration_secs': 0.1209} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1479.997752] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-60c93e4e-db83-4dde-ab91-d7984de84f9a tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281660', 'volume_id': 'fdc9ac56-be2c-4fc5-a830-1f781e497d60', 'name': 'volume-fdc9ac56-be2c-4fc5-a830-1f781e497d60', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '812b40f7-3b7e-43a5-92ae-df3891e74e4c', 'attached_at': '', 'detached_at': '', 'volume_id': 'fdc9ac56-be2c-4fc5-a830-1f781e497d60', 'serial': 'fdc9ac56-be2c-4fc5-a830-1f781e497d60'} {{(pid=61868) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1481.031969] env[61868]: DEBUG nova.objects.instance [None req-60c93e4e-db83-4dde-ab91-d7984de84f9a tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lazy-loading 'flavor' on Instance uuid 812b40f7-3b7e-43a5-92ae-df3891e74e4c {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1481.538525] env[61868]: DEBUG oslo_concurrency.lockutils [None req-60c93e4e-db83-4dde-ab91-d7984de84f9a tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lock "812b40f7-3b7e-43a5-92ae-df3891e74e4c" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.229s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1481.696908] env[61868]: INFO nova.compute.manager [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Rescuing [ 1481.697206] env[61868]: DEBUG oslo_concurrency.lockutils [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquiring lock "refresh_cache-812b40f7-3b7e-43a5-92ae-df3891e74e4c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1481.697365] env[61868]: DEBUG oslo_concurrency.lockutils [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquired lock "refresh_cache-812b40f7-3b7e-43a5-92ae-df3891e74e4c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1481.697542] env[61868]: DEBUG nova.network.neutron [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1482.406896] env[61868]: DEBUG nova.network.neutron [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Updating instance_info_cache with network_info: [{"id": "5aa4db9f-d6ec-47fa-b327-8ab288ed9255", "address": "fa:16:3e:23:f9:38", "network": {"id": "5be7f7b7-0d32-48b1-979d-5a4a872e5b1a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1791355381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.201", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b3904d578a0e47d58eae75b1d4fbf41c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5aa4db9f-d6", "ovs_interfaceid": "5aa4db9f-d6ec-47fa-b327-8ab288ed9255", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1482.910158] env[61868]: DEBUG oslo_concurrency.lockutils [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Releasing lock "refresh_cache-812b40f7-3b7e-43a5-92ae-df3891e74e4c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1483.440861] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1483.441169] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ad67d19b-b489-4501-8f60-ca44d65f5e04 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1483.449210] env[61868]: DEBUG oslo_vmware.api [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1483.449210] env[61868]: value = "task-1316234" [ 1483.449210] env[61868]: _type = "Task" [ 1483.449210] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1483.458300] env[61868]: DEBUG oslo_vmware.api [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316234, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1483.959624] env[61868]: DEBUG oslo_vmware.api [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316234, 'name': PowerOffVM_Task, 'duration_secs': 0.172601} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1483.960759] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1483.960934] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b8ab524-d01a-4bac-b2ca-c552eb64e559 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1483.981908] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae6bb1ed-f30c-4417-9e0d-85a36b66654c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1484.006938] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1484.007213] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8567242b-fd3e-4841-be37-16f458cf0f0d {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1484.014584] env[61868]: DEBUG oslo_vmware.api [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1484.014584] env[61868]: value = "task-1316235" [ 1484.014584] env[61868]: _type = "Task" [ 1484.014584] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1484.025174] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] VM already powered off {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1484.025382] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Processing image c4fdfa42-c9a0-41ca-801e-4e55c3d4328d {{(pid=61868) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1484.025629] env[61868]: DEBUG oslo_concurrency.lockutils [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1484.025781] env[61868]: DEBUG oslo_concurrency.lockutils [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1484.025960] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1484.026201] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c332c7d7-5b07-4ad0-a0ee-cf0c8c4c2670 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1484.034016] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61868) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1484.034211] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61868) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1484.034868] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4206c0b0-7359-4d1f-9ef8-3eda9cf6cfc7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1484.039565] env[61868]: DEBUG oslo_vmware.api [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1484.039565] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5294f679-5153-c194-bf6d-d7d9a882c52b" [ 1484.039565] env[61868]: _type = "Task" [ 1484.039565] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1484.046260] env[61868]: DEBUG oslo_vmware.api [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5294f679-5153-c194-bf6d-d7d9a882c52b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1484.550099] env[61868]: DEBUG oslo_vmware.api [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]5294f679-5153-c194-bf6d-d7d9a882c52b, 'name': SearchDatastore_Task, 'duration_secs': 0.008181} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1484.550906] env[61868]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b97137c-57f8-4636-b908-c05ecae57c74 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1484.555722] env[61868]: DEBUG oslo_vmware.api [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1484.555722] env[61868]: value = "session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e51d96-20ef-6266-ea1d-be732fd95ed0" [ 1484.555722] env[61868]: _type = "Task" [ 1484.555722] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1484.563198] env[61868]: DEBUG oslo_vmware.api [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e51d96-20ef-6266-ea1d-be732fd95ed0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1485.065902] env[61868]: DEBUG oslo_vmware.api [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': session[52fffe61-622b-7fcf-6f02-284bb4d393a6]52e51d96-20ef-6266-ea1d-be732fd95ed0, 'name': SearchDatastore_Task, 'duration_secs': 0.009067} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1485.066187] env[61868]: DEBUG oslo_concurrency.lockutils [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1485.066461] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 812b40f7-3b7e-43a5-92ae-df3891e74e4c/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d-rescue.vmdk. {{(pid=61868) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1485.066721] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1a86b22b-79f7-45e1-aebe-a599ac8ea6d3 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1485.073421] env[61868]: DEBUG oslo_vmware.api [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1485.073421] env[61868]: value = "task-1316236" [ 1485.073421] env[61868]: _type = "Task" [ 1485.073421] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1485.080503] env[61868]: DEBUG oslo_vmware.api [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316236, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1485.582805] env[61868]: DEBUG oslo_vmware.api [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316236, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.417132} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1485.583230] env[61868]: INFO nova.virt.vmwareapi.ds_util [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d.vmdk to [datastore1] 812b40f7-3b7e-43a5-92ae-df3891e74e4c/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d-rescue.vmdk. [ 1485.583853] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbae8632-fbc7-4060-adb6-274cc96039be {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1485.610364] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Reconfiguring VM instance instance-00000072 to attach disk [datastore1] 812b40f7-3b7e-43a5-92ae-df3891e74e4c/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d-rescue.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1485.610596] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9ce3538f-8a7b-49c7-a194-87a2889408cb {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1485.628093] env[61868]: DEBUG oslo_vmware.api [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1485.628093] env[61868]: value = "task-1316237" [ 1485.628093] env[61868]: _type = "Task" [ 1485.628093] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1485.635365] env[61868]: DEBUG oslo_vmware.api [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316237, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1486.137535] env[61868]: DEBUG oslo_vmware.api [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316237, 'name': ReconfigVM_Task, 'duration_secs': 0.302579} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1486.137823] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Reconfigured VM instance instance-00000072 to attach disk [datastore1] 812b40f7-3b7e-43a5-92ae-df3891e74e4c/c4fdfa42-c9a0-41ca-801e-4e55c3d4328d-rescue.vmdk or device None with type thin {{(pid=61868) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1486.138696] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38e922dd-8050-42d9-93d1-d184b1523d5b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1486.165619] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b29d718f-c6b7-4a2d-acff-20b43912ea71 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1486.180733] env[61868]: DEBUG oslo_vmware.api [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1486.180733] env[61868]: value = "task-1316238" [ 1486.180733] env[61868]: _type = "Task" [ 1486.180733] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1486.187969] env[61868]: DEBUG oslo_vmware.api [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316238, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1486.690479] env[61868]: DEBUG oslo_vmware.api [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316238, 'name': ReconfigVM_Task, 'duration_secs': 0.282059} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1486.690841] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1486.691020] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7dce471f-1346-45be-b2f3-75bf64712289 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1486.697125] env[61868]: DEBUG oslo_vmware.api [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1486.697125] env[61868]: value = "task-1316239" [ 1486.697125] env[61868]: _type = "Task" [ 1486.697125] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1486.704201] env[61868]: DEBUG oslo_vmware.api [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316239, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1487.206663] env[61868]: DEBUG oslo_vmware.api [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316239, 'name': PowerOnVM_Task, 'duration_secs': 0.364414} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1487.206895] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1487.209660] env[61868]: DEBUG nova.compute.manager [None req-980caa35-c10b-48dc-954e-0436a25b5388 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1487.210457] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79ab0014-b877-4dde-b90e-9c85da68c667 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1488.576424] env[61868]: INFO nova.compute.manager [None req-21a585a2-e0f7-4b76-940c-8bc5517eacd1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Unrescuing [ 1488.576806] env[61868]: DEBUG oslo_concurrency.lockutils [None req-21a585a2-e0f7-4b76-940c-8bc5517eacd1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquiring lock "refresh_cache-812b40f7-3b7e-43a5-92ae-df3891e74e4c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1488.576909] env[61868]: DEBUG oslo_concurrency.lockutils [None req-21a585a2-e0f7-4b76-940c-8bc5517eacd1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquired lock "refresh_cache-812b40f7-3b7e-43a5-92ae-df3891e74e4c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1488.577072] env[61868]: DEBUG nova.network.neutron [None req-21a585a2-e0f7-4b76-940c-8bc5517eacd1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Building network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1489.290436] env[61868]: DEBUG nova.network.neutron [None req-21a585a2-e0f7-4b76-940c-8bc5517eacd1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Updating instance_info_cache with network_info: [{"id": "5aa4db9f-d6ec-47fa-b327-8ab288ed9255", "address": "fa:16:3e:23:f9:38", "network": {"id": "5be7f7b7-0d32-48b1-979d-5a4a872e5b1a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1791355381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.201", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b3904d578a0e47d58eae75b1d4fbf41c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5aa4db9f-d6", "ovs_interfaceid": "5aa4db9f-d6ec-47fa-b327-8ab288ed9255", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1489.792924] env[61868]: DEBUG oslo_concurrency.lockutils [None req-21a585a2-e0f7-4b76-940c-8bc5517eacd1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Releasing lock "refresh_cache-812b40f7-3b7e-43a5-92ae-df3891e74e4c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1489.793628] env[61868]: DEBUG nova.objects.instance [None req-21a585a2-e0f7-4b76-940c-8bc5517eacd1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lazy-loading 'flavor' on Instance uuid 812b40f7-3b7e-43a5-92ae-df3891e74e4c {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1490.299316] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d42e941c-1f36-4a09-b2f3-dd462462c9c6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1490.323762] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-21a585a2-e0f7-4b76-940c-8bc5517eacd1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1490.323762] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-12ea28e6-f03f-4afd-a2ca-b14a13c85279 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1490.329915] env[61868]: DEBUG oslo_vmware.api [None req-21a585a2-e0f7-4b76-940c-8bc5517eacd1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1490.329915] env[61868]: value = "task-1316240" [ 1490.329915] env[61868]: _type = "Task" [ 1490.329915] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1490.337129] env[61868]: DEBUG oslo_vmware.api [None req-21a585a2-e0f7-4b76-940c-8bc5517eacd1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316240, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1490.840081] env[61868]: DEBUG oslo_vmware.api [None req-21a585a2-e0f7-4b76-940c-8bc5517eacd1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316240, 'name': PowerOffVM_Task, 'duration_secs': 0.206015} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1490.840462] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-21a585a2-e0f7-4b76-940c-8bc5517eacd1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1490.845573] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-21a585a2-e0f7-4b76-940c-8bc5517eacd1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Reconfiguring VM instance instance-00000072 to detach disk 2002 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1490.845851] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cf6d8634-242e-44db-9078-1fc0542a3488 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1490.863363] env[61868]: DEBUG oslo_vmware.api [None req-21a585a2-e0f7-4b76-940c-8bc5517eacd1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1490.863363] env[61868]: value = "task-1316241" [ 1490.863363] env[61868]: _type = "Task" [ 1490.863363] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1490.870465] env[61868]: DEBUG oslo_vmware.api [None req-21a585a2-e0f7-4b76-940c-8bc5517eacd1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316241, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1491.374226] env[61868]: DEBUG oslo_vmware.api [None req-21a585a2-e0f7-4b76-940c-8bc5517eacd1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316241, 'name': ReconfigVM_Task, 'duration_secs': 0.219187} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1491.374226] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-21a585a2-e0f7-4b76-940c-8bc5517eacd1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Reconfigured VM instance instance-00000072 to detach disk 2002 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1491.374226] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-21a585a2-e0f7-4b76-940c-8bc5517eacd1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Powering on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1491.374398] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e87777bb-952f-446e-b882-e1d37bec324a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1491.381063] env[61868]: DEBUG oslo_vmware.api [None req-21a585a2-e0f7-4b76-940c-8bc5517eacd1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1491.381063] env[61868]: value = "task-1316242" [ 1491.381063] env[61868]: _type = "Task" [ 1491.381063] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1491.387598] env[61868]: DEBUG oslo_vmware.api [None req-21a585a2-e0f7-4b76-940c-8bc5517eacd1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316242, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1491.890646] env[61868]: DEBUG oslo_vmware.api [None req-21a585a2-e0f7-4b76-940c-8bc5517eacd1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316242, 'name': PowerOnVM_Task, 'duration_secs': 0.357434} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1491.891046] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-21a585a2-e0f7-4b76-940c-8bc5517eacd1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Powered on the VM {{(pid=61868) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1491.891238] env[61868]: DEBUG nova.compute.manager [None req-21a585a2-e0f7-4b76-940c-8bc5517eacd1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Checking state {{(pid=61868) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1491.892020] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-771524f7-0d03-45a1-aa24-596fbda26956 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1493.045072] env[61868]: DEBUG nova.compute.manager [req-52c8397b-05ba-4fe1-b973-7ffc984f0a5f req-f9827c4d-7129-45f3-8b5d-11f58257ad43 service nova] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Received event network-changed-5aa4db9f-d6ec-47fa-b327-8ab288ed9255 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1493.045320] env[61868]: DEBUG nova.compute.manager [req-52c8397b-05ba-4fe1-b973-7ffc984f0a5f req-f9827c4d-7129-45f3-8b5d-11f58257ad43 service nova] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Refreshing instance network info cache due to event network-changed-5aa4db9f-d6ec-47fa-b327-8ab288ed9255. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1493.045517] env[61868]: DEBUG oslo_concurrency.lockutils [req-52c8397b-05ba-4fe1-b973-7ffc984f0a5f req-f9827c4d-7129-45f3-8b5d-11f58257ad43 service nova] Acquiring lock "refresh_cache-812b40f7-3b7e-43a5-92ae-df3891e74e4c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1493.045646] env[61868]: DEBUG oslo_concurrency.lockutils [req-52c8397b-05ba-4fe1-b973-7ffc984f0a5f req-f9827c4d-7129-45f3-8b5d-11f58257ad43 service nova] Acquired lock "refresh_cache-812b40f7-3b7e-43a5-92ae-df3891e74e4c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1493.045810] env[61868]: DEBUG nova.network.neutron [req-52c8397b-05ba-4fe1-b973-7ffc984f0a5f req-f9827c4d-7129-45f3-8b5d-11f58257ad43 service nova] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Refreshing network info cache for port 5aa4db9f-d6ec-47fa-b327-8ab288ed9255 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1493.762039] env[61868]: DEBUG nova.network.neutron [req-52c8397b-05ba-4fe1-b973-7ffc984f0a5f req-f9827c4d-7129-45f3-8b5d-11f58257ad43 service nova] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Updated VIF entry in instance network info cache for port 5aa4db9f-d6ec-47fa-b327-8ab288ed9255. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1493.762461] env[61868]: DEBUG nova.network.neutron [req-52c8397b-05ba-4fe1-b973-7ffc984f0a5f req-f9827c4d-7129-45f3-8b5d-11f58257ad43 service nova] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Updating instance_info_cache with network_info: [{"id": "5aa4db9f-d6ec-47fa-b327-8ab288ed9255", "address": "fa:16:3e:23:f9:38", "network": {"id": "5be7f7b7-0d32-48b1-979d-5a4a872e5b1a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1791355381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.201", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b3904d578a0e47d58eae75b1d4fbf41c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5aa4db9f-d6", "ovs_interfaceid": "5aa4db9f-d6ec-47fa-b327-8ab288ed9255", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1494.265765] env[61868]: DEBUG oslo_concurrency.lockutils [req-52c8397b-05ba-4fe1-b973-7ffc984f0a5f req-f9827c4d-7129-45f3-8b5d-11f58257ad43 service nova] Releasing lock "refresh_cache-812b40f7-3b7e-43a5-92ae-df3891e74e4c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1495.080365] env[61868]: DEBUG nova.compute.manager [req-a67740f6-665d-4737-bb2f-aea0eb974925 req-66cc7b85-2c75-4d5a-b3f7-fe7ea72437ad service nova] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Received event network-changed-5aa4db9f-d6ec-47fa-b327-8ab288ed9255 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1495.080581] env[61868]: DEBUG nova.compute.manager [req-a67740f6-665d-4737-bb2f-aea0eb974925 req-66cc7b85-2c75-4d5a-b3f7-fe7ea72437ad service nova] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Refreshing instance network info cache due to event network-changed-5aa4db9f-d6ec-47fa-b327-8ab288ed9255. {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1495.080825] env[61868]: DEBUG oslo_concurrency.lockutils [req-a67740f6-665d-4737-bb2f-aea0eb974925 req-66cc7b85-2c75-4d5a-b3f7-fe7ea72437ad service nova] Acquiring lock "refresh_cache-812b40f7-3b7e-43a5-92ae-df3891e74e4c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1495.080983] env[61868]: DEBUG oslo_concurrency.lockutils [req-a67740f6-665d-4737-bb2f-aea0eb974925 req-66cc7b85-2c75-4d5a-b3f7-fe7ea72437ad service nova] Acquired lock "refresh_cache-812b40f7-3b7e-43a5-92ae-df3891e74e4c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1495.081171] env[61868]: DEBUG nova.network.neutron [req-a67740f6-665d-4737-bb2f-aea0eb974925 req-66cc7b85-2c75-4d5a-b3f7-fe7ea72437ad service nova] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Refreshing network info cache for port 5aa4db9f-d6ec-47fa-b327-8ab288ed9255 {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1495.774760] env[61868]: DEBUG nova.network.neutron [req-a67740f6-665d-4737-bb2f-aea0eb974925 req-66cc7b85-2c75-4d5a-b3f7-fe7ea72437ad service nova] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Updated VIF entry in instance network info cache for port 5aa4db9f-d6ec-47fa-b327-8ab288ed9255. {{(pid=61868) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1495.775161] env[61868]: DEBUG nova.network.neutron [req-a67740f6-665d-4737-bb2f-aea0eb974925 req-66cc7b85-2c75-4d5a-b3f7-fe7ea72437ad service nova] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Updating instance_info_cache with network_info: [{"id": "5aa4db9f-d6ec-47fa-b327-8ab288ed9255", "address": "fa:16:3e:23:f9:38", "network": {"id": "5be7f7b7-0d32-48b1-979d-5a4a872e5b1a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1791355381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.201", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b3904d578a0e47d58eae75b1d4fbf41c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5aa4db9f-d6", "ovs_interfaceid": "5aa4db9f-d6ec-47fa-b327-8ab288ed9255", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1496.278291] env[61868]: DEBUG oslo_concurrency.lockutils [req-a67740f6-665d-4737-bb2f-aea0eb974925 req-66cc7b85-2c75-4d5a-b3f7-fe7ea72437ad service nova] Releasing lock "refresh_cache-812b40f7-3b7e-43a5-92ae-df3891e74e4c" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1518.076818] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1519.075105] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1519.075105] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1519.075105] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61868) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1519.075345] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager.update_available_resource {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1519.578713] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1519.579118] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1519.579204] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1519.579314] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61868) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1519.580308] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e31d29d-ecc6-4068-adc8-61e9484a7e7c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1519.588781] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-303d4508-64b7-4630-8295-11ec808801c0 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1519.602348] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efa6bbc1-ad73-4b1b-9440-ef938bbfd339 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1519.608261] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-badb84d7-b908-4b0a-8206-5b5e1adc5006 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1519.637074] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181114MB free_disk=134GB free_vcpus=48 pci_devices=None {{(pid=61868) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1519.637223] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1519.637399] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1520.664030] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 28a14ebd-6003-4e20-b556-2a3b939f15c8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1520.664358] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 6ce00281-a201-445a-bb75-aac07fd481aa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1520.664358] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Instance 812b40f7-3b7e-43a5-92ae-df3891e74e4c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61868) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1520.664433] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=61868) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1520.664525] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=61868) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1520.709522] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0afd829-c13e-4f99-90af-e8eb873b8829 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1520.716782] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5913e5e6-c855-41f3-964f-6cb647956543 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1520.745130] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eca37d67-073b-440e-9bac-616817907a79 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1520.752104] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40777309-fbab-4d9d-9de6-ec44b6c9a856 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1520.764550] env[61868]: DEBUG nova.compute.provider_tree [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1521.267818] env[61868]: DEBUG nova.scheduler.client.report [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1521.269212] env[61868]: DEBUG nova.compute.resource_tracker [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61868) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1521.269396] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.632s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1523.826892] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3bbc1c0b-cfb5-47d3-94f9-83d9037cd2a1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquiring lock "812b40f7-3b7e-43a5-92ae-df3891e74e4c" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1523.827285] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3bbc1c0b-cfb5-47d3-94f9-83d9037cd2a1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lock "812b40f7-3b7e-43a5-92ae-df3891e74e4c" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1524.265140] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1524.265378] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1524.265526] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Starting heal instance info cache {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1524.330106] env[61868]: INFO nova.compute.manager [None req-3bbc1c0b-cfb5-47d3-94f9-83d9037cd2a1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Detaching volume fdc9ac56-be2c-4fc5-a830-1f781e497d60 [ 1524.359938] env[61868]: INFO nova.virt.block_device [None req-3bbc1c0b-cfb5-47d3-94f9-83d9037cd2a1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Attempting to driver detach volume fdc9ac56-be2c-4fc5-a830-1f781e497d60 from mountpoint /dev/sdb [ 1524.360210] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bbc1c0b-cfb5-47d3-94f9-83d9037cd2a1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Volume detach. Driver type: vmdk {{(pid=61868) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1524.360433] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bbc1c0b-cfb5-47d3-94f9-83d9037cd2a1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281660', 'volume_id': 'fdc9ac56-be2c-4fc5-a830-1f781e497d60', 'name': 'volume-fdc9ac56-be2c-4fc5-a830-1f781e497d60', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '812b40f7-3b7e-43a5-92ae-df3891e74e4c', 'attached_at': '', 'detached_at': '', 'volume_id': 'fdc9ac56-be2c-4fc5-a830-1f781e497d60', 'serial': 'fdc9ac56-be2c-4fc5-a830-1f781e497d60'} {{(pid=61868) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1524.361282] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25a75243-f55f-4db8-9cc0-3fb1e239a861 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1524.382306] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-babf0132-51aa-4a7b-9184-dde63cfd6db7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1524.388942] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d90f9c88-bf36-464f-9c95-380d4cc89156 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1524.407938] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ac2ccba-5ae4-47a2-a2bc-9d63397ac65b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1524.421765] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bbc1c0b-cfb5-47d3-94f9-83d9037cd2a1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] The volume has not been displaced from its original location: [datastore1] volume-fdc9ac56-be2c-4fc5-a830-1f781e497d60/volume-fdc9ac56-be2c-4fc5-a830-1f781e497d60.vmdk. No consolidation needed. {{(pid=61868) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1524.426820] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bbc1c0b-cfb5-47d3-94f9-83d9037cd2a1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Reconfiguring VM instance instance-00000072 to detach disk 2001 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1524.427104] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b47ee9e2-eccc-4775-a149-a767942785c8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1524.444649] env[61868]: DEBUG oslo_vmware.api [None req-3bbc1c0b-cfb5-47d3-94f9-83d9037cd2a1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1524.444649] env[61868]: value = "task-1316243" [ 1524.444649] env[61868]: _type = "Task" [ 1524.444649] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1524.453284] env[61868]: DEBUG oslo_vmware.api [None req-3bbc1c0b-cfb5-47d3-94f9-83d9037cd2a1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316243, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1524.801739] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquiring lock "refresh_cache-6ce00281-a201-445a-bb75-aac07fd481aa" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1524.801895] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Acquired lock "refresh_cache-6ce00281-a201-445a-bb75-aac07fd481aa" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1524.802052] env[61868]: DEBUG nova.network.neutron [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Forcefully refreshing network info cache for instance {{(pid=61868) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1524.954145] env[61868]: DEBUG oslo_vmware.api [None req-3bbc1c0b-cfb5-47d3-94f9-83d9037cd2a1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316243, 'name': ReconfigVM_Task, 'duration_secs': 0.209315} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1524.954495] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bbc1c0b-cfb5-47d3-94f9-83d9037cd2a1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Reconfigured VM instance instance-00000072 to detach disk 2001 {{(pid=61868) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1524.958960] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d8ad331e-4e5a-4a03-aefc-f06febd8c2fd {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1524.973298] env[61868]: DEBUG oslo_vmware.api [None req-3bbc1c0b-cfb5-47d3-94f9-83d9037cd2a1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1524.973298] env[61868]: value = "task-1316244" [ 1524.973298] env[61868]: _type = "Task" [ 1524.973298] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1524.981051] env[61868]: DEBUG oslo_vmware.api [None req-3bbc1c0b-cfb5-47d3-94f9-83d9037cd2a1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316244, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1525.483543] env[61868]: DEBUG oslo_vmware.api [None req-3bbc1c0b-cfb5-47d3-94f9-83d9037cd2a1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316244, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1525.983512] env[61868]: DEBUG oslo_vmware.api [None req-3bbc1c0b-cfb5-47d3-94f9-83d9037cd2a1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316244, 'name': ReconfigVM_Task, 'duration_secs': 0.745404} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1525.983848] env[61868]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bbc1c0b-cfb5-47d3-94f9-83d9037cd2a1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-281660', 'volume_id': 'fdc9ac56-be2c-4fc5-a830-1f781e497d60', 'name': 'volume-fdc9ac56-be2c-4fc5-a830-1f781e497d60', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '812b40f7-3b7e-43a5-92ae-df3891e74e4c', 'attached_at': '', 'detached_at': '', 'volume_id': 'fdc9ac56-be2c-4fc5-a830-1f781e497d60', 'serial': 'fdc9ac56-be2c-4fc5-a830-1f781e497d60'} {{(pid=61868) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1526.003832] env[61868]: DEBUG nova.network.neutron [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Updating instance_info_cache with network_info: [{"id": "dc8823b5-592e-40dc-81dd-5bad3cf56bce", "address": "fa:16:3e:c5:24:eb", "network": {"id": "5be7f7b7-0d32-48b1-979d-5a4a872e5b1a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1791355381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b3904d578a0e47d58eae75b1d4fbf41c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc8823b5-59", "ovs_interfaceid": "dc8823b5-592e-40dc-81dd-5bad3cf56bce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1526.506445] env[61868]: DEBUG oslo_concurrency.lockutils [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Releasing lock "refresh_cache-6ce00281-a201-445a-bb75-aac07fd481aa" {{(pid=61868) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1526.506712] env[61868]: DEBUG nova.compute.manager [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Updated the network info_cache for instance {{(pid=61868) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1526.506928] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1526.507131] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1526.522878] env[61868]: DEBUG nova.objects.instance [None req-3bbc1c0b-cfb5-47d3-94f9-83d9037cd2a1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lazy-loading 'flavor' on Instance uuid 812b40f7-3b7e-43a5-92ae-df3891e74e4c {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1527.531880] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3bbc1c0b-cfb5-47d3-94f9-83d9037cd2a1 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lock "812b40f7-3b7e-43a5-92ae-df3891e74e4c" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.704s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1528.075540] env[61868]: DEBUG oslo_service.periodic_task [None req-bea220a1-553c-44e0-a68e-efd918df27aa None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61868) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1528.127381] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ac7bcaea-99da-4886-a06c-6f510b13dba0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquiring lock "812b40f7-3b7e-43a5-92ae-df3891e74e4c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1528.127629] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ac7bcaea-99da-4886-a06c-6f510b13dba0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lock "812b40f7-3b7e-43a5-92ae-df3891e74e4c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1528.127926] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ac7bcaea-99da-4886-a06c-6f510b13dba0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquiring lock "812b40f7-3b7e-43a5-92ae-df3891e74e4c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1528.128196] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ac7bcaea-99da-4886-a06c-6f510b13dba0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lock "812b40f7-3b7e-43a5-92ae-df3891e74e4c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1528.128415] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ac7bcaea-99da-4886-a06c-6f510b13dba0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lock "812b40f7-3b7e-43a5-92ae-df3891e74e4c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1528.130650] env[61868]: INFO nova.compute.manager [None req-ac7bcaea-99da-4886-a06c-6f510b13dba0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Terminating instance [ 1528.132405] env[61868]: DEBUG nova.compute.manager [None req-ac7bcaea-99da-4886-a06c-6f510b13dba0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1528.132643] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7bcaea-99da-4886-a06c-6f510b13dba0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1528.133501] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6887bbe8-b0f6-44a8-852d-fa63cf2a88f5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1528.141494] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac7bcaea-99da-4886-a06c-6f510b13dba0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1528.141779] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6f5e9132-0027-4a54-bd27-1bc1e7bceba4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1528.148525] env[61868]: DEBUG oslo_vmware.api [None req-ac7bcaea-99da-4886-a06c-6f510b13dba0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1528.148525] env[61868]: value = "task-1316245" [ 1528.148525] env[61868]: _type = "Task" [ 1528.148525] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1528.158253] env[61868]: DEBUG oslo_vmware.api [None req-ac7bcaea-99da-4886-a06c-6f510b13dba0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316245, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1528.659078] env[61868]: DEBUG oslo_vmware.api [None req-ac7bcaea-99da-4886-a06c-6f510b13dba0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316245, 'name': PowerOffVM_Task, 'duration_secs': 0.173932} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1528.659459] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac7bcaea-99da-4886-a06c-6f510b13dba0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1528.659565] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7bcaea-99da-4886-a06c-6f510b13dba0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1528.659913] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4b7b17ad-b562-4f94-b886-404cc51dce02 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1528.723956] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7bcaea-99da-4886-a06c-6f510b13dba0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1528.724269] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7bcaea-99da-4886-a06c-6f510b13dba0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Deleting contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1528.724513] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac7bcaea-99da-4886-a06c-6f510b13dba0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Deleting the datastore file [datastore1] 812b40f7-3b7e-43a5-92ae-df3891e74e4c {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1528.724815] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ead4f6a8-e5c6-4b71-b732-41520543dc40 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1528.730722] env[61868]: DEBUG oslo_vmware.api [None req-ac7bcaea-99da-4886-a06c-6f510b13dba0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1528.730722] env[61868]: value = "task-1316247" [ 1528.730722] env[61868]: _type = "Task" [ 1528.730722] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1528.738139] env[61868]: DEBUG oslo_vmware.api [None req-ac7bcaea-99da-4886-a06c-6f510b13dba0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316247, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1529.240749] env[61868]: DEBUG oslo_vmware.api [None req-ac7bcaea-99da-4886-a06c-6f510b13dba0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316247, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.163495} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1529.241089] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac7bcaea-99da-4886-a06c-6f510b13dba0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1529.241327] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7bcaea-99da-4886-a06c-6f510b13dba0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Deleted contents of the VM from datastore datastore1 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1529.241555] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7bcaea-99da-4886-a06c-6f510b13dba0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1529.241782] env[61868]: INFO nova.compute.manager [None req-ac7bcaea-99da-4886-a06c-6f510b13dba0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1529.242048] env[61868]: DEBUG oslo.service.loopingcall [None req-ac7bcaea-99da-4886-a06c-6f510b13dba0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1529.242247] env[61868]: DEBUG nova.compute.manager [-] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1529.242344] env[61868]: DEBUG nova.network.neutron [-] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1529.767528] env[61868]: DEBUG nova.compute.manager [req-0f31a9f7-2504-41a0-b52b-32710052c588 req-8478cbf7-2ca0-4215-afa8-36d8ef612ef6 service nova] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Received event network-vif-deleted-5aa4db9f-d6ec-47fa-b327-8ab288ed9255 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1529.767807] env[61868]: INFO nova.compute.manager [req-0f31a9f7-2504-41a0-b52b-32710052c588 req-8478cbf7-2ca0-4215-afa8-36d8ef612ef6 service nova] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Neutron deleted interface 5aa4db9f-d6ec-47fa-b327-8ab288ed9255; detaching it from the instance and deleting it from the info cache [ 1529.767952] env[61868]: DEBUG nova.network.neutron [req-0f31a9f7-2504-41a0-b52b-32710052c588 req-8478cbf7-2ca0-4215-afa8-36d8ef612ef6 service nova] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1530.213269] env[61868]: DEBUG nova.network.neutron [-] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1530.270608] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a0404a5f-3691-43d8-b962-dedf1078cd22 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1530.280286] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9825a413-813c-4fcf-8114-24279302f6fc {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1530.303289] env[61868]: DEBUG nova.compute.manager [req-0f31a9f7-2504-41a0-b52b-32710052c588 req-8478cbf7-2ca0-4215-afa8-36d8ef612ef6 service nova] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Detach interface failed, port_id=5aa4db9f-d6ec-47fa-b327-8ab288ed9255, reason: Instance 812b40f7-3b7e-43a5-92ae-df3891e74e4c could not be found. {{(pid=61868) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1530.716475] env[61868]: INFO nova.compute.manager [-] [instance: 812b40f7-3b7e-43a5-92ae-df3891e74e4c] Took 1.47 seconds to deallocate network for instance. [ 1531.223112] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ac7bcaea-99da-4886-a06c-6f510b13dba0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1531.223112] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ac7bcaea-99da-4886-a06c-6f510b13dba0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1531.223112] env[61868]: DEBUG nova.objects.instance [None req-ac7bcaea-99da-4886-a06c-6f510b13dba0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lazy-loading 'resources' on Instance uuid 812b40f7-3b7e-43a5-92ae-df3891e74e4c {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1531.779935] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba986ede-f72a-40e6-8e4b-510403e25642 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1531.787229] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8b64a3b-0b3c-4b3a-a8fb-fc00019c15e8 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1531.816559] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f18d3b1b-20d1-4175-859e-27c9da6b986f {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1531.822943] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f863c60-75e6-4d80-b5b8-0cd0d2051051 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1531.835302] env[61868]: DEBUG nova.compute.provider_tree [None req-ac7bcaea-99da-4886-a06c-6f510b13dba0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1532.338941] env[61868]: DEBUG nova.scheduler.client.report [None req-ac7bcaea-99da-4886-a06c-6f510b13dba0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1532.843983] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ac7bcaea-99da-4886-a06c-6f510b13dba0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.621s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1532.865626] env[61868]: INFO nova.scheduler.client.report [None req-ac7bcaea-99da-4886-a06c-6f510b13dba0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Deleted allocations for instance 812b40f7-3b7e-43a5-92ae-df3891e74e4c [ 1533.373639] env[61868]: DEBUG oslo_concurrency.lockutils [None req-ac7bcaea-99da-4886-a06c-6f510b13dba0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lock "812b40f7-3b7e-43a5-92ae-df3891e74e4c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.246s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1538.147543] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c39b9475-bb9a-4ff0-829e-9c25465641f0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquiring lock "6ce00281-a201-445a-bb75-aac07fd481aa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1538.147939] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c39b9475-bb9a-4ff0-829e-9c25465641f0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lock "6ce00281-a201-445a-bb75-aac07fd481aa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1538.148052] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c39b9475-bb9a-4ff0-829e-9c25465641f0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquiring lock "6ce00281-a201-445a-bb75-aac07fd481aa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1538.148230] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c39b9475-bb9a-4ff0-829e-9c25465641f0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lock "6ce00281-a201-445a-bb75-aac07fd481aa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1538.148404] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c39b9475-bb9a-4ff0-829e-9c25465641f0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lock "6ce00281-a201-445a-bb75-aac07fd481aa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1538.150731] env[61868]: INFO nova.compute.manager [None req-c39b9475-bb9a-4ff0-829e-9c25465641f0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Terminating instance [ 1538.152517] env[61868]: DEBUG nova.compute.manager [None req-c39b9475-bb9a-4ff0-829e-9c25465641f0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1538.152711] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c39b9475-bb9a-4ff0-829e-9c25465641f0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1538.153555] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-372a2243-a1f3-4b01-838c-6c7ddad91932 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1538.160962] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c39b9475-bb9a-4ff0-829e-9c25465641f0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1538.161476] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f066b038-2986-4fd1-84f1-fb80d30f4fe6 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1538.167994] env[61868]: DEBUG oslo_vmware.api [None req-c39b9475-bb9a-4ff0-829e-9c25465641f0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1538.167994] env[61868]: value = "task-1316249" [ 1538.167994] env[61868]: _type = "Task" [ 1538.167994] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1538.176625] env[61868]: DEBUG oslo_vmware.api [None req-c39b9475-bb9a-4ff0-829e-9c25465641f0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316249, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1538.678148] env[61868]: DEBUG oslo_vmware.api [None req-c39b9475-bb9a-4ff0-829e-9c25465641f0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316249, 'name': PowerOffVM_Task, 'duration_secs': 0.213955} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1538.678422] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-c39b9475-bb9a-4ff0-829e-9c25465641f0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1538.678601] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c39b9475-bb9a-4ff0-829e-9c25465641f0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1538.678882] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5da39f54-fdec-48e3-8573-61596909fb96 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1538.739533] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c39b9475-bb9a-4ff0-829e-9c25465641f0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1538.739766] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c39b9475-bb9a-4ff0-829e-9c25465641f0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Deleting contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1538.740084] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-c39b9475-bb9a-4ff0-829e-9c25465641f0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Deleting the datastore file [datastore2] 6ce00281-a201-445a-bb75-aac07fd481aa {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1538.740388] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-61de47ff-6ee9-478a-a764-f2a83f48eba4 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1538.746116] env[61868]: DEBUG oslo_vmware.api [None req-c39b9475-bb9a-4ff0-829e-9c25465641f0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1538.746116] env[61868]: value = "task-1316251" [ 1538.746116] env[61868]: _type = "Task" [ 1538.746116] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1538.753544] env[61868]: DEBUG oslo_vmware.api [None req-c39b9475-bb9a-4ff0-829e-9c25465641f0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316251, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1539.256191] env[61868]: DEBUG oslo_vmware.api [None req-c39b9475-bb9a-4ff0-829e-9c25465641f0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316251, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.164325} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1539.256559] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-c39b9475-bb9a-4ff0-829e-9c25465641f0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1539.256625] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c39b9475-bb9a-4ff0-829e-9c25465641f0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Deleted contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1539.256790] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-c39b9475-bb9a-4ff0-829e-9c25465641f0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1539.256972] env[61868]: INFO nova.compute.manager [None req-c39b9475-bb9a-4ff0-829e-9c25465641f0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1539.257239] env[61868]: DEBUG oslo.service.loopingcall [None req-c39b9475-bb9a-4ff0-829e-9c25465641f0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1539.257441] env[61868]: DEBUG nova.compute.manager [-] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1539.257538] env[61868]: DEBUG nova.network.neutron [-] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1539.507522] env[61868]: DEBUG nova.compute.manager [req-b963700c-c6cc-4425-bb9e-6222ded4492d req-2a9aa86a-1721-4c8d-8b8a-856fcaaa867d service nova] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Received event network-vif-deleted-dc8823b5-592e-40dc-81dd-5bad3cf56bce {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1539.507711] env[61868]: INFO nova.compute.manager [req-b963700c-c6cc-4425-bb9e-6222ded4492d req-2a9aa86a-1721-4c8d-8b8a-856fcaaa867d service nova] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Neutron deleted interface dc8823b5-592e-40dc-81dd-5bad3cf56bce; detaching it from the instance and deleting it from the info cache [ 1539.507917] env[61868]: DEBUG nova.network.neutron [req-b963700c-c6cc-4425-bb9e-6222ded4492d req-2a9aa86a-1721-4c8d-8b8a-856fcaaa867d service nova] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1539.985732] env[61868]: DEBUG nova.network.neutron [-] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1540.010708] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d79bee67-6d9b-4b79-90ef-93470b2ee635 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1540.019854] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41216e24-8463-4770-ac23-b7d9683472df {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1540.042136] env[61868]: DEBUG nova.compute.manager [req-b963700c-c6cc-4425-bb9e-6222ded4492d req-2a9aa86a-1721-4c8d-8b8a-856fcaaa867d service nova] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Detach interface failed, port_id=dc8823b5-592e-40dc-81dd-5bad3cf56bce, reason: Instance 6ce00281-a201-445a-bb75-aac07fd481aa could not be found. {{(pid=61868) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1540.488632] env[61868]: INFO nova.compute.manager [-] [instance: 6ce00281-a201-445a-bb75-aac07fd481aa] Took 1.23 seconds to deallocate network for instance. [ 1540.996798] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c39b9475-bb9a-4ff0-829e-9c25465641f0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1540.997121] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c39b9475-bb9a-4ff0-829e-9c25465641f0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1540.997361] env[61868]: DEBUG nova.objects.instance [None req-c39b9475-bb9a-4ff0-829e-9c25465641f0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lazy-loading 'resources' on Instance uuid 6ce00281-a201-445a-bb75-aac07fd481aa {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1541.544931] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5109d31e-6903-4792-971c-4bd48e0514f1 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1541.552581] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-714e9a28-bb85-4b79-94f9-b70733abf67c {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1541.582471] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39a021d6-c7b2-4bfc-b545-d8e5072c787b {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1541.589146] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26759ef7-8cac-4527-bdda-c6db8d3f06b5 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1541.601412] env[61868]: DEBUG nova.compute.provider_tree [None req-c39b9475-bb9a-4ff0-829e-9c25465641f0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1542.104702] env[61868]: DEBUG nova.scheduler.client.report [None req-c39b9475-bb9a-4ff0-829e-9c25465641f0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1542.609793] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c39b9475-bb9a-4ff0-829e-9c25465641f0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.612s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1542.636159] env[61868]: INFO nova.scheduler.client.report [None req-c39b9475-bb9a-4ff0-829e-9c25465641f0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Deleted allocations for instance 6ce00281-a201-445a-bb75-aac07fd481aa [ 1543.143967] env[61868]: DEBUG oslo_concurrency.lockutils [None req-c39b9475-bb9a-4ff0-829e-9c25465641f0 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lock "6ce00281-a201-445a-bb75-aac07fd481aa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.996s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1543.663648] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3bd1f9c9-8bb5-4797-9199-92148b24e352 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquiring lock "28a14ebd-6003-4e20-b556-2a3b939f15c8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1543.664058] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3bd1f9c9-8bb5-4797-9199-92148b24e352 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lock "28a14ebd-6003-4e20-b556-2a3b939f15c8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1543.664160] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3bd1f9c9-8bb5-4797-9199-92148b24e352 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquiring lock "28a14ebd-6003-4e20-b556-2a3b939f15c8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1543.664374] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3bd1f9c9-8bb5-4797-9199-92148b24e352 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lock "28a14ebd-6003-4e20-b556-2a3b939f15c8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1543.664602] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3bd1f9c9-8bb5-4797-9199-92148b24e352 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lock "28a14ebd-6003-4e20-b556-2a3b939f15c8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1543.666825] env[61868]: INFO nova.compute.manager [None req-3bd1f9c9-8bb5-4797-9199-92148b24e352 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Terminating instance [ 1543.668651] env[61868]: DEBUG nova.compute.manager [None req-3bd1f9c9-8bb5-4797-9199-92148b24e352 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Start destroying the instance on the hypervisor. {{(pid=61868) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1543.668888] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-3bd1f9c9-8bb5-4797-9199-92148b24e352 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Destroying instance {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1543.669736] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d33299a-33e9-4838-9e3f-d7a445774bc9 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1543.677289] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bd1f9c9-8bb5-4797-9199-92148b24e352 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Powering off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1543.677530] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-957b63ea-9de5-40f9-8bc2-ababbe30233a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1543.683997] env[61868]: DEBUG oslo_vmware.api [None req-3bd1f9c9-8bb5-4797-9199-92148b24e352 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1543.683997] env[61868]: value = "task-1316252" [ 1543.683997] env[61868]: _type = "Task" [ 1543.683997] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1543.691838] env[61868]: DEBUG oslo_vmware.api [None req-3bd1f9c9-8bb5-4797-9199-92148b24e352 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316252, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1544.195352] env[61868]: DEBUG oslo_vmware.api [None req-3bd1f9c9-8bb5-4797-9199-92148b24e352 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316252, 'name': PowerOffVM_Task, 'duration_secs': 0.212791} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1544.195619] env[61868]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bd1f9c9-8bb5-4797-9199-92148b24e352 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Powered off the VM {{(pid=61868) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1544.195787] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-3bd1f9c9-8bb5-4797-9199-92148b24e352 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Unregistering the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1544.196040] env[61868]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6539e69d-b8e0-4bd9-84db-0da34a43b3d7 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1544.256122] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-3bd1f9c9-8bb5-4797-9199-92148b24e352 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Unregistered the VM {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1544.256359] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-3bd1f9c9-8bb5-4797-9199-92148b24e352 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Deleting contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1544.256539] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-3bd1f9c9-8bb5-4797-9199-92148b24e352 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Deleting the datastore file [datastore2] 28a14ebd-6003-4e20-b556-2a3b939f15c8 {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1544.256801] env[61868]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3ad7815b-c755-498d-ad66-d983a9345f56 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1544.265145] env[61868]: DEBUG oslo_vmware.api [None req-3bd1f9c9-8bb5-4797-9199-92148b24e352 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for the task: (returnval){ [ 1544.265145] env[61868]: value = "task-1316254" [ 1544.265145] env[61868]: _type = "Task" [ 1544.265145] env[61868]: } to complete. {{(pid=61868) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1544.273085] env[61868]: DEBUG oslo_vmware.api [None req-3bd1f9c9-8bb5-4797-9199-92148b24e352 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316254, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1544.775071] env[61868]: DEBUG oslo_vmware.api [None req-3bd1f9c9-8bb5-4797-9199-92148b24e352 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Task: {'id': task-1316254, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.135048} completed successfully. {{(pid=61868) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1544.775493] env[61868]: DEBUG nova.virt.vmwareapi.ds_util [None req-3bd1f9c9-8bb5-4797-9199-92148b24e352 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Deleted the datastore file {{(pid=61868) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1544.775493] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-3bd1f9c9-8bb5-4797-9199-92148b24e352 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Deleted contents of the VM from datastore datastore2 {{(pid=61868) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1544.775493] env[61868]: DEBUG nova.virt.vmwareapi.vmops [None req-3bd1f9c9-8bb5-4797-9199-92148b24e352 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Instance destroyed {{(pid=61868) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1544.775644] env[61868]: INFO nova.compute.manager [None req-3bd1f9c9-8bb5-4797-9199-92148b24e352 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1544.775883] env[61868]: DEBUG oslo.service.loopingcall [None req-3bd1f9c9-8bb5-4797-9199-92148b24e352 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61868) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1544.776091] env[61868]: DEBUG nova.compute.manager [-] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Deallocating network for instance {{(pid=61868) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1544.776187] env[61868]: DEBUG nova.network.neutron [-] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] deallocate_for_instance() {{(pid=61868) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1545.033221] env[61868]: DEBUG nova.compute.manager [req-c337b1db-adfb-4d50-a336-82b04e10a20f req-c5dd389e-cf74-42b5-a805-aa2cb1e09aa8 service nova] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Received event network-vif-deleted-b1bf2dec-5f14-461f-aa54-f6d67bc0dcb5 {{(pid=61868) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1545.033378] env[61868]: INFO nova.compute.manager [req-c337b1db-adfb-4d50-a336-82b04e10a20f req-c5dd389e-cf74-42b5-a805-aa2cb1e09aa8 service nova] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Neutron deleted interface b1bf2dec-5f14-461f-aa54-f6d67bc0dcb5; detaching it from the instance and deleting it from the info cache [ 1545.033564] env[61868]: DEBUG nova.network.neutron [req-c337b1db-adfb-4d50-a336-82b04e10a20f req-c5dd389e-cf74-42b5-a805-aa2cb1e09aa8 service nova] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1545.514143] env[61868]: DEBUG nova.network.neutron [-] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Updating instance_info_cache with network_info: [] {{(pid=61868) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1545.535730] env[61868]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-33a1eadf-4092-4176-a6cf-8b3ad6e38e0a {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1545.546345] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4826875-0dfc-4a4b-8f5f-7eb97a97af49 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1545.567846] env[61868]: DEBUG nova.compute.manager [req-c337b1db-adfb-4d50-a336-82b04e10a20f req-c5dd389e-cf74-42b5-a805-aa2cb1e09aa8 service nova] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Detach interface failed, port_id=b1bf2dec-5f14-461f-aa54-f6d67bc0dcb5, reason: Instance 28a14ebd-6003-4e20-b556-2a3b939f15c8 could not be found. {{(pid=61868) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1546.017149] env[61868]: INFO nova.compute.manager [-] [instance: 28a14ebd-6003-4e20-b556-2a3b939f15c8] Took 1.24 seconds to deallocate network for instance. [ 1546.523993] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3bd1f9c9-8bb5-4797-9199-92148b24e352 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1546.524360] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3bd1f9c9-8bb5-4797-9199-92148b24e352 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1546.524622] env[61868]: DEBUG nova.objects.instance [None req-3bd1f9c9-8bb5-4797-9199-92148b24e352 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lazy-loading 'resources' on Instance uuid 28a14ebd-6003-4e20-b556-2a3b939f15c8 {{(pid=61868) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1547.058165] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f50fdeed-01b7-472d-8313-f27daa023345 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1547.065818] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2600d465-3697-4574-9711-b0bf3adb4904 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1547.095471] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a1d8da7-67de-4639-9c44-e14e0cb0f7ce {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1547.102095] env[61868]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94945a9a-61cd-4080-8e71-3313f6171476 {{(pid=61868) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1547.114956] env[61868]: DEBUG nova.compute.provider_tree [None req-3bd1f9c9-8bb5-4797-9199-92148b24e352 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Inventory has not changed in ProviderTree for provider: 83e67721-2ac9-4a23-aa31-82aca86979c8 {{(pid=61868) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1547.618244] env[61868]: DEBUG nova.scheduler.client.report [None req-3bd1f9c9-8bb5-4797-9199-92148b24e352 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Inventory has not changed for provider 83e67721-2ac9-4a23-aa31-82aca86979c8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 134, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61868) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1548.123025] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3bd1f9c9-8bb5-4797-9199-92148b24e352 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.598s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1548.150853] env[61868]: INFO nova.scheduler.client.report [None req-3bd1f9c9-8bb5-4797-9199-92148b24e352 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Deleted allocations for instance 28a14ebd-6003-4e20-b556-2a3b939f15c8 [ 1548.659078] env[61868]: DEBUG oslo_concurrency.lockutils [None req-3bd1f9c9-8bb5-4797-9199-92148b24e352 tempest-ServerRescueNegativeTestJSON-886592011 tempest-ServerRescueNegativeTestJSON-886592011-project-member] Lock "28a14ebd-6003-4e20-b556-2a3b939f15c8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.995s {{(pid=61868) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}}